Mercurial > vec
diff include/vec/impl/generic.h @ 37:4b5a557aa64f
*: turns out extern is a practical joke. rewrite to be always inline again
the sample benchmark performs about 3x as well with optimizations
disabled :)
author | Paper <paper@tflc.us> |
---|---|
date | Sat, 26 Apr 2025 01:04:35 -0400 |
parents | 677c03c382b8 |
children | fd42f9b1b95e |
line wrap: on
line diff
--- a/include/vec/impl/generic.h Fri Apr 25 17:40:55 2025 -0400 +++ b/include/vec/impl/generic.h Sat Apr 26 01:04:35 2025 -0400 @@ -22,7 +22,8 @@ * SOFTWARE. **/ -/* Generic array-based implementation. */ +/* This file is automatically generated! Do not edit it directly! + * Edit the code that generates it in utils/gengeneric.c --paper */ #ifndef VEC_IMPL_GENERIC_H_ #define VEC_IMPL_GENERIC_H_ @@ -31,110 +32,4526 @@ // ----------------------------------------------------------------- -// TODO implement these so we don't waste stack space by doing the -// fallbacks -#define VEC_GENERIC_DEFINE_OPERATIONS_SIGN(sign, csign, bits, size) \ - static v##sign##int##bits##x##size v##sign##int##bits##x##size##_generic_load_aligned(const vec_##sign##int##bits in[size]) \ +#define VEC_GENERIC_OPERATION(op, sign, bits, size) \ + do { \ + int i; \ + \ + for (i = 0; i < size; i++) \ + vec1.generic[i] = (op); \ + \ + return vec1; \ + } while (0) + +#define VEC_GENERIC_BUILTIN_OPERATION(op, sign, bits, size) \ + VEC_GENERIC_OPERATION(vec1.generic[i] op vec2.generic[i], sign, bits, size) + +#define VEC_GENERIC_CMP(op, sign, bits, size) \ + VEC_GENERIC_OPERATION((vec1.generic[i] op vec2.generic[i]) ? (vec_##sign##int##bits)VEC_MAX_OF_TYPE(vec_uint##bits) : 0, sign, bits, size) + +/* okay, now we can do this crap: */ + +#define VEC_GENERIC_SPLAT(sign, bits, size) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_splat(vec_##sign##int##bits x) \ + { \ + v##sign##int##bits##x##size vec; \ + for (int i = 0; i < size; i++) \ + vec.generic[i] = x; \ + return vec; \ + } + +#define VEC_GENERIC_LOAD_EX(name, sign, bits, size) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_##name(const vec_##sign##int##bits in[size]) \ { \ v##sign##int##bits##x##size vec; \ - memcpy(vec.generic, in, sizeof(vec_##sign##int##bits) * size); \ + memcpy(&vec, in, sizeof(vec_##sign##int##bits) * size); \ return vec; \ - } \ - \ - static void v##sign##int##bits##x##size##_generic_store_aligned(v##sign##int##bits##x##size vec, vec_##sign##int##bits out[size]) \ + } + +#define VEC_GENERIC_LOAD_ALIGNED(sign, bits, size) VEC_GENERIC_LOAD_EX(load_aligned, sign, bits, size) +#define VEC_GENERIC_LOAD(sign, bits, size) VEC_GENERIC_LOAD_EX(load, sign, bits, size) + +#define VEC_GENERIC_STORE_EX(name, sign, bits, size) \ + VEC_FUNC_IMPL void v##sign##int##bits##x##size##_##name(v##sign##int##bits##x##size vec, vec_##sign##int##bits out[size]) \ + { \ + memcpy(out, &vec, sizeof(vec_##sign##int##bits) * size); \ + } + +#define VEC_GENERIC_STORE_ALIGNED(sign, bits, size) VEC_GENERIC_STORE_EX(store_aligned, sign, bits, size) +#define VEC_GENERIC_STORE(sign, bits, size) VEC_GENERIC_STORE_EX(store, sign, bits, size) + +#define VEC_GENERIC_ADD(sign, bits, size) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_add(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ { \ - memcpy(out, vec.generic, sizeof(vec_##sign##int##bits) * size); \ - } \ + VEC_GENERIC_BUILTIN_OPERATION(+, sign, bits, size); \ + } + +#define VEC_GENERIC_SUB(sign, bits, size) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_sub(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ + { \ + VEC_GENERIC_BUILTIN_OPERATION(-, sign, bits, size); \ + } + +#define VEC_GENERIC_MUL(sign, bits, size) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_mul(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ + { \ + VEC_GENERIC_BUILTIN_OPERATION(*, sign, bits, size); \ + } + +#define VEC_GENERIC_DIV(sign, bits, size) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_div(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ + { \ + VEC_GENERIC_OPERATION(vec2.generic[i] ? (vec1.generic[i] / vec2.generic[i]) : 0, sign, bits, size); \ + } + +#define VEC_GENERIC_AVG(sign, bits, size) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_avg(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ + { \ + for (int i = 0; i < size; i++) \ + vec1.generic[i] = vec_##sign##avg(vec1.generic[i], vec2.generic[i]); \ \ - static v##sign##int##bits##x##size##_impl v##sign##int##bits##x##size##_impl_generic = { \ - /* .splat = */ NULL, \ - v##sign##int##bits##x##size##_generic_load_aligned, \ - v##sign##int##bits##x##size##_generic_load_aligned, \ - v##sign##int##bits##x##size##_generic_store_aligned, \ - v##sign##int##bits##x##size##_generic_store_aligned, \ - }; - -#define VEC_GENERIC_DEFINE_OPERATIONS(bits, size) \ - VEC_GENERIC_DEFINE_OPERATIONS_SIGN( , , bits, size) \ - VEC_GENERIC_DEFINE_OPERATIONS_SIGN(u, U, bits, size) - -VEC_GENERIC_DEFINE_OPERATIONS(8, 2) -VEC_GENERIC_DEFINE_OPERATIONS(16, 2) -VEC_GENERIC_DEFINE_OPERATIONS(32, 2) -VEC_GENERIC_DEFINE_OPERATIONS(64, 2) - -#undef VEC_GENERIC_DEFINE_OPERATIONS -#undef VEC_GENERIC_DEFINE_OPERATIONS_SIGN - -// ----------------------------------------------------------------- -// now we can just keep doubling the same implementation - -#define VEC_GENERIC_DEFINE_OPERATIONS_SIGN(sign, csign, bits, size, halfsize) \ - static v##sign##int##bits##x##size v##sign##int##bits##x##size##_generic_load_aligned(const vec_##sign##int##bits in[size]) \ + return vec1; \ + } + +#define VEC_GENERIC_AND(sign, bits, size) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_and(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ + { \ + VEC_GENERIC_BUILTIN_OPERATION(&, sign, bits, size); \ + } + +#define VEC_GENERIC_OR(sign, bits, size) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_or(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ + { \ + VEC_GENERIC_BUILTIN_OPERATION(|, sign, bits, size); \ + } + +#define VEC_GENERIC_XOR(sign, bits, size) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_xor(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ + { \ + VEC_GENERIC_BUILTIN_OPERATION(^, sign, bits, size); \ + } + +#define VEC_GENERIC_NOT(sign, bits, size) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_not(v##sign##int##bits##x##size vec) \ + { \ + return v##sign##int##bits##x##size##_xor(vec, v##sign##int##bits##x##size##_splat((vec_##sign##int##bits)VEC_MAX_OF_TYPE(vec_uint##bits))); \ + } + +#define VEC_GENERIC_CMPLT(sign, bits, size) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_cmplt(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ + { \ + VEC_GENERIC_CMP(<, sign, bits, size); \ + } + +#define VEC_GENERIC_CMPLE(sign, bits, size) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_cmple(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ + { \ + return v##sign##int##bits##x##size##_not(v##sign##int##bits##x##size##_cmpgt(vec1, vec2)); \ + } + +#define VEC_GENERIC_CMPEQ(sign, bits, size) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_cmpeq(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ + { \ + VEC_GENERIC_CMP(==, sign, bits, size); \ + } + +#define VEC_GENERIC_CMPGE(sign, bits, size) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_cmpge(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ + { \ + return v##sign##int##bits##x##size##_not(v##sign##int##bits##x##size##_cmplt(vec1, vec2)); \ + } + +#define VEC_GENERIC_CMPGT(sign, bits, size) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_cmpgt(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ + { \ + VEC_GENERIC_CMP(>, sign, bits, size); \ + } + +#define VEC_GENERIC_LSHIFT(sign, bits, size) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_lshift(v##sign##int##bits##x##size vec1, vuint##bits##x##size vec2) \ + { \ + VEC_GENERIC_OPERATION(vec_##sign##lshift(vec1.generic[i], vec2.generic[i]), sign, bits, size); \ + } + +#define VEC_GENERIC_RSHIFT(sign, bits, size) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_rshift(v##sign##int##bits##x##size vec1, vuint##bits##x##size vec2) \ + { \ + VEC_GENERIC_OPERATION(vec_##sign##rshift(vec1.generic[i], vec2.generic[i]), sign, bits, size); \ + } + +#define VEC_GENERIC_LRSHIFT(sign, bits, size) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_lrshift(v##sign##int##bits##x##size vec1, vuint##bits##x##size vec2) \ + { \ + VEC_GENERIC_OPERATION(vec_urshift((vec_uint##bits)vec1.generic[i], vec2.generic[i]), sign, bits, size); \ + } + +#define VEC_GENERIC_MIN(sign, bits, size) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_min(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ + { \ + v##sign##int##bits##x##size cmplt = v##sign##int##bits##x##size##_cmplt(vec1, vec2); \ + \ + v##sign##int##bits##x##size a = v##sign##int##bits##x##size##_and(vec1, cmplt); \ + v##sign##int##bits##x##size b = v##sign##int##bits##x##size##_and(vec2, v##sign##int##bits##x##size##_not(cmplt)); \ + \ + return v##sign##int##bits##x##size##_or(a, b); \ + } + +#define VEC_GENERIC_MAX(sign, bits, size) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_max(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ + { \ + v##sign##int##bits##x##size cmplt = v##sign##int##bits##x##size##_cmpgt(vec1, vec2); \ + \ + v##sign##int##bits##x##size a = v##sign##int##bits##x##size##_and(vec1, cmplt); \ + v##sign##int##bits##x##size b = v##sign##int##bits##x##size##_and(vec2, v##sign##int##bits##x##size##_not(cmplt)); \ + \ + return v##sign##int##bits##x##size##_or(a, b); \ + } + +#define VEC_GENERIC_DBL_SPLAT(sign, bits, size, halfsize) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_splat(vec_##sign##int##bits x) \ { \ v##sign##int##bits##x##size vec; \ - vec.generic[0] = v##sign##int##bits##x##halfsize##_load_aligned(in); \ - vec.generic[1] = v##sign##int##bits##x##halfsize##_load_aligned(in + halfsize); \ + \ + vec.generic[0] = v##sign##int##bits##x##halfsize##_splat(x); \ + vec.generic[1] = v##sign##int##bits##x##halfsize##_splat(x); \ + \ return vec; \ - } \ - \ - static v##sign##int##bits##x##size v##sign##int##bits##x##size##_generic_load(const vec_##sign##int##bits in[size]) \ + } + +#define VEC_GENERIC_DBL_LOAD_EX(name, sign, bits, size, halfsize) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_##name(const vec_##sign##int##bits x[size]) \ { \ v##sign##int##bits##x##size vec; \ - vec.generic[0] = v##sign##int##bits##x##halfsize##_load(in); \ - vec.generic[1] = v##sign##int##bits##x##halfsize##_load(in + halfsize); \ + \ + vec.generic[0] = v##sign##int##bits##x##halfsize##_##name(x); \ + vec.generic[1] = v##sign##int##bits##x##halfsize##_##name(x + halfsize); \ + \ return vec; \ - } \ - \ - static void v##sign##int##bits##x##size##_generic_store_aligned(v##sign##int##bits##x##size vec, vec_##sign##int##bits out[size]) \ + } + +#define VEC_GENERIC_DBL_LOAD(sign, bits, size, halfsize) VEC_GENERIC_DBL_LOAD_EX(load, sign, bits, size, halfsize) +#define VEC_GENERIC_DBL_LOAD_ALIGNED(sign, bits, size, halfsize) VEC_GENERIC_DBL_LOAD_EX(load_aligned, sign, bits, size, halfsize) + +#define VEC_GENERIC_DBL_STORE_EX(name, sign, bits, size, halfsize) \ + VEC_FUNC_IMPL void v##sign##int##bits##x##size##_##name(v##sign##int##bits##x##size vec, vec_##sign##int##bits x[size]) \ + { \ + v##sign##int##bits##x##halfsize##_##name(vec.generic[0], x); \ + v##sign##int##bits##x##halfsize##_##name(vec.generic[1], x + halfsize); \ + } + +#define VEC_GENERIC_DBL_STORE(sign, bits, size, halfsize) VEC_GENERIC_DBL_STORE_EX(store, sign, bits, size, halfsize) +#define VEC_GENERIC_DBL_STORE_ALIGNED(sign, bits, size, halfsize) VEC_GENERIC_DBL_STORE_EX(store_aligned, sign, bits, size, halfsize) + +#define VEC_GENERIC_DBL_OP(name, sign, bits, size, halfsize, secondsign) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_##name(v##sign##int##bits##x##size vec1, v##secondsign##int##bits##x##size vec2) \ { \ - v##sign##int##bits##x##halfsize##_store_aligned(vec.generic[0], out); \ - v##sign##int##bits##x##halfsize##_store_aligned(vec.generic[1], out + halfsize); \ - } \ + vec1.generic[0] = v##sign##int##bits##x##halfsize##_##name(vec1.generic[0], vec2.generic[0]); \ + vec1.generic[1] = v##sign##int##bits##x##halfsize##_##name(vec1.generic[1], vec2.generic[1]); \ \ - static void v##sign##int##bits##x##size##_generic_store(v##sign##int##bits##x##size vec, vec_##sign##int##bits out[size]) \ + return vec1; \ + } + +#define VEC_GENERIC_DBL_ADD(sign, bits, size, halfsize) VEC_GENERIC_DBL_OP(add, sign, bits, size, halfsize, sign) +#define VEC_GENERIC_DBL_SUB(sign, bits, size, halfsize) VEC_GENERIC_DBL_OP(sub, sign, bits, size, halfsize, sign) +#define VEC_GENERIC_DBL_MUL(sign, bits, size, halfsize) VEC_GENERIC_DBL_OP(mul, sign, bits, size, halfsize, sign) +#define VEC_GENERIC_DBL_DIV(sign, bits, size, halfsize) VEC_GENERIC_DBL_OP(div, sign, bits, size, halfsize, sign) +#define VEC_GENERIC_DBL_AVG(sign, bits, size, halfsize) VEC_GENERIC_DBL_OP(avg, sign, bits, size, halfsize, sign) +#define VEC_GENERIC_DBL_LSHIFT(sign, bits, size, halfsize) VEC_GENERIC_DBL_OP(lshift, sign, bits, size, halfsize, u) +#define VEC_GENERIC_DBL_RSHIFT(sign, bits, size, halfsize) VEC_GENERIC_DBL_OP(rshift, sign, bits, size, halfsize, u) +#define VEC_GENERIC_DBL_LRSHIFT(sign, bits, size, halfsize) VEC_GENERIC_DBL_OP(lrshift, sign, bits, size, halfsize, u) +#define VEC_GENERIC_DBL_AND(sign, bits, size, halfsize) VEC_GENERIC_DBL_OP(and, sign, bits, size, halfsize, sign) +#define VEC_GENERIC_DBL_OR(sign, bits, size, halfsize) VEC_GENERIC_DBL_OP(or, sign, bits, size, halfsize, sign) +#define VEC_GENERIC_DBL_XOR(sign, bits, size, halfsize) VEC_GENERIC_DBL_OP(xor, sign, bits, size, halfsize, sign) +#define VEC_GENERIC_DBL_MIN(sign, bits, size, halfsize) VEC_GENERIC_DBL_OP(min, sign, bits, size, halfsize, sign) +#define VEC_GENERIC_DBL_MAX(sign, bits, size, halfsize) VEC_GENERIC_DBL_OP(max, sign, bits, size, halfsize, sign) +#define VEC_GENERIC_DBL_CMPLT(sign, bits, size, halfsize) VEC_GENERIC_DBL_OP(cmplt, sign, bits, size, halfsize, sign) +#define VEC_GENERIC_DBL_CMPLE(sign, bits, size, halfsize) VEC_GENERIC_DBL_OP(cmple, sign, bits, size, halfsize, sign) +#define VEC_GENERIC_DBL_CMPEQ(sign, bits, size, halfsize) VEC_GENERIC_DBL_OP(cmpeq, sign, bits, size, halfsize, sign) +#define VEC_GENERIC_DBL_CMPGE(sign, bits, size, halfsize) VEC_GENERIC_DBL_OP(cmpge, sign, bits, size, halfsize, sign) +#define VEC_GENERIC_DBL_CMPGT(sign, bits, size, halfsize) VEC_GENERIC_DBL_OP(cmpgt, sign, bits, size, halfsize, sign) + +#define VEC_GENERIC_DBL_NOT(sign, bits, size, halfsize) \ + VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_not(v##sign##int##bits##x##size vec) \ { \ - v##sign##int##bits##x##halfsize##_store(vec.generic[0], out); \ - v##sign##int##bits##x##halfsize##_store(vec.generic[1], out + halfsize); \ - } \ + vec.generic[0] = v##sign##int##bits##x##halfsize##_not(vec.generic[0]); \ + vec.generic[1] = v##sign##int##bits##x##halfsize##_not(vec.generic[1]); \ \ - static v##sign##int##bits##x##size##_impl v##sign##int##bits##x##size##_impl_generic = { \ - /* .splat = */ NULL, \ - v##sign##int##bits##x##size##_generic_load_aligned, \ - v##sign##int##bits##x##size##_generic_load, \ - v##sign##int##bits##x##size##_generic_store_aligned, \ - v##sign##int##bits##x##size##_generic_store, \ - }; - -#define VEC_GENERIC_DEFINE_OPERATIONS(bits, size, halfsize) \ - VEC_GENERIC_DEFINE_OPERATIONS_SIGN( , , bits, size, halfsize) \ - VEC_GENERIC_DEFINE_OPERATIONS_SIGN(u, U, bits, size, halfsize) - -// 32-bit -VEC_GENERIC_DEFINE_OPERATIONS(8, 4, 2) - -// 64-bit -VEC_GENERIC_DEFINE_OPERATIONS(8, 8, 4) -VEC_GENERIC_DEFINE_OPERATIONS(16, 4, 2) - -// 128-bit -VEC_GENERIC_DEFINE_OPERATIONS(8, 16, 8) -VEC_GENERIC_DEFINE_OPERATIONS(16, 8, 4) -VEC_GENERIC_DEFINE_OPERATIONS(32, 4, 2) - -// 256-bit -VEC_GENERIC_DEFINE_OPERATIONS(8, 32, 16) -VEC_GENERIC_DEFINE_OPERATIONS(16, 16, 8) -VEC_GENERIC_DEFINE_OPERATIONS(32, 8, 4) -VEC_GENERIC_DEFINE_OPERATIONS(64, 4, 2) - -// 512-bit -VEC_GENERIC_DEFINE_OPERATIONS(8, 64, 32) -VEC_GENERIC_DEFINE_OPERATIONS(16, 32, 16) -VEC_GENERIC_DEFINE_OPERATIONS(32, 16, 8) -VEC_GENERIC_DEFINE_OPERATIONS(64, 8, 4) - -#undef VEC_GENERIC_DEFINE_OPERATIONS -#undef VEC_GENERIC_DEFINE_OPERATIONS_SIGN + return vec; \ + } + +/* ------------------------------------------------------------------------ */ +/* PREPROCESSOR HELL INCOMING */ + + + +/* vuint8x2 */ + +#ifndef VINT8x2_SPLAT_DEFINED +VEC_GENERIC_SPLAT(/* nothing */, 8, 2) +# define VINT8x2_SPLAT_DEFINED +#endif +#ifndef VINT8x2_LOAD_ALIGNED_DEFINED +VEC_GENERIC_LOAD_ALIGNED(/* nothing */, 8, 2) +# define VINT8x2_LOAD_ALIGNED_DEFINED +#endif +#ifndef VINT8x2_LOAD_DEFINED +VEC_GENERIC_LOAD(/* nothing */, 8, 2) +# define VINT8x2_LOAD_DEFINED +#endif +#ifndef VINT8x2_STORE_ALIGNED_DEFINED +VEC_GENERIC_STORE_ALIGNED(/* nothing */, 8, 2) +# define VINT8x2_STORE_ALIGNED_DEFINED +#endif +#ifndef VINT8x2_STORE_DEFINED +VEC_GENERIC_STORE(/* nothing */, 8, 2) +# define VINT8x2_STORE_DEFINED +#endif +#ifndef VINT8x2_ADD_DEFINED +VEC_GENERIC_ADD(/* nothing */, 8, 2) +# define VINT8x2_ADD_DEFINED +#endif +#ifndef VINT8x2_SUB_DEFINED +VEC_GENERIC_SUB(/* nothing */, 8, 2) +# define VINT8x2_SUB_DEFINED +#endif +#ifndef VINT8x2_MUL_DEFINED +VEC_GENERIC_MUL(/* nothing */, 8, 2) +# define VINT8x2_MUL_DEFINED +#endif +#ifndef VINT8x2_DIV_DEFINED +VEC_GENERIC_DIV(/* nothing */, 8, 2) +# define VINT8x2_DIV_DEFINED +#endif +#ifndef VINT8x2_AVG_DEFINED +VEC_GENERIC_AVG(/* nothing */, 8, 2) +# define VINT8x2_AVG_DEFINED +#endif +#ifndef VINT8x2_AND_DEFINED +VEC_GENERIC_AND(/* nothing */, 8, 2) +# define VINT8x2_AND_DEFINED +#endif +#ifndef VINT8x2_OR_DEFINED +VEC_GENERIC_OR(/* nothing */, 8, 2) +# define VINT8x2_OR_DEFINED +#endif +#ifndef VINT8x2_XOR_DEFINED +VEC_GENERIC_XOR(/* nothing */, 8, 2) +# define VINT8x2_XOR_DEFINED +#endif +#ifndef VINT8x2_NOT_DEFINED +VEC_GENERIC_NOT(/* nothing */, 8, 2) +# define VINT8x2_NOT_DEFINED +#endif +#ifndef VINT8x2_CMPLT_DEFINED +VEC_GENERIC_CMPLT(/* nothing */, 8, 2) +# define VINT8x2_CMPLT_DEFINED +#endif +#ifndef VINT8x2_CMPEQ_DEFINED +VEC_GENERIC_CMPEQ(/* nothing */, 8, 2) +# define VINT8x2_CMPEQ_DEFINED +#endif +#ifndef VINT8x2_CMPGT_DEFINED +VEC_GENERIC_CMPGT(/* nothing */, 8, 2) +# define VINT8x2_CMPGT_DEFINED +#endif +#ifndef VINT8x2_CMPLE_DEFINED +VEC_GENERIC_CMPLE(/* nothing */, 8, 2) +# define VINT8x2_CMPLE_DEFINED +#endif +#ifndef VINT8x2_CMPGE_DEFINED +VEC_GENERIC_CMPGE(/* nothing */, 8, 2) +# define VINT8x2_CMPGE_DEFINED +#endif +#ifndef VINT8x2_MIN_DEFINED +VEC_GENERIC_MIN(/* nothing */, 8, 2) +# define VINT8x2_MIN_DEFINED +#endif +#ifndef VINT8x2_MAX_DEFINED +VEC_GENERIC_MAX(/* nothing */, 8, 2) +# define VINT8x2_MAX_DEFINED +#endif +#ifndef VINT8x2_RSHIFT_DEFINED +VEC_GENERIC_RSHIFT(/* nothing */, 8, 2) +# define VINT8x2_RSHIFT_DEFINED +#endif +#ifndef VINT8x2_LRSHIFT_DEFINED +VEC_GENERIC_LRSHIFT(/* nothing */, 8, 2) +# define VINT8x2_LRSHIFT_DEFINED +#endif +#ifndef VINT8x2_LSHIFT_DEFINED +VEC_GENERIC_LSHIFT(/* nothing */, 8, 2) +# define VINT8x2_LSHIFT_DEFINED +#endif + + +/* vint8x2 */ + +#ifndef VUINT8x2_SPLAT_DEFINED +VEC_GENERIC_SPLAT(u, 8, 2) +# define VUINT8x2_SPLAT_DEFINED +#endif +#ifndef VUINT8x2_LOAD_ALIGNED_DEFINED +VEC_GENERIC_LOAD_ALIGNED(u, 8, 2) +# define VUINT8x2_LOAD_ALIGNED_DEFINED +#endif +#ifndef VUINT8x2_LOAD_DEFINED +VEC_GENERIC_LOAD(u, 8, 2) +# define VUINT8x2_LOAD_DEFINED +#endif +#ifndef VUINT8x2_STORE_ALIGNED_DEFINED +VEC_GENERIC_STORE_ALIGNED(u, 8, 2) +# define VUINT8x2_STORE_ALIGNED_DEFINED +#endif +#ifndef VUINT8x2_STORE_DEFINED +VEC_GENERIC_STORE(u, 8, 2) +# define VUINT8x2_STORE_DEFINED +#endif +#ifndef VUINT8x2_ADD_DEFINED +VEC_GENERIC_ADD(u, 8, 2) +# define VUINT8x2_ADD_DEFINED +#endif +#ifndef VUINT8x2_SUB_DEFINED +VEC_GENERIC_SUB(u, 8, 2) +# define VUINT8x2_SUB_DEFINED +#endif +#ifndef VUINT8x2_MUL_DEFINED +VEC_GENERIC_MUL(u, 8, 2) +# define VUINT8x2_MUL_DEFINED +#endif +#ifndef VUINT8x2_DIV_DEFINED +VEC_GENERIC_DIV(u, 8, 2) +# define VUINT8x2_DIV_DEFINED +#endif +#ifndef VUINT8x2_AVG_DEFINED +VEC_GENERIC_AVG(u, 8, 2) +# define VUINT8x2_AVG_DEFINED +#endif +#ifndef VUINT8x2_AND_DEFINED +VEC_GENERIC_AND(u, 8, 2) +# define VUINT8x2_AND_DEFINED +#endif +#ifndef VUINT8x2_OR_DEFINED +VEC_GENERIC_OR(u, 8, 2) +# define VUINT8x2_OR_DEFINED +#endif +#ifndef VUINT8x2_XOR_DEFINED +VEC_GENERIC_XOR(u, 8, 2) +# define VUINT8x2_XOR_DEFINED +#endif +#ifndef VUINT8x2_NOT_DEFINED +VEC_GENERIC_NOT(u, 8, 2) +# define VUINT8x2_NOT_DEFINED +#endif +#ifndef VUINT8x2_CMPLT_DEFINED +VEC_GENERIC_CMPLT(u, 8, 2) +# define VUINT8x2_CMPLT_DEFINED +#endif +#ifndef VUINT8x2_CMPEQ_DEFINED +VEC_GENERIC_CMPEQ(u, 8, 2) +# define VUINT8x2_CMPEQ_DEFINED +#endif +#ifndef VUINT8x2_CMPGT_DEFINED +VEC_GENERIC_CMPGT(u, 8, 2) +# define VUINT8x2_CMPGT_DEFINED +#endif +#ifndef VUINT8x2_CMPLE_DEFINED +VEC_GENERIC_CMPLE(u, 8, 2) +# define VUINT8x2_CMPLE_DEFINED +#endif +#ifndef VUINT8x2_CMPGE_DEFINED +VEC_GENERIC_CMPGE(u, 8, 2) +# define VUINT8x2_CMPGE_DEFINED +#endif +#ifndef VUINT8x2_MIN_DEFINED +VEC_GENERIC_MIN(u, 8, 2) +# define VUINT8x2_MIN_DEFINED +#endif +#ifndef VUINT8x2_MAX_DEFINED +VEC_GENERIC_MAX(u, 8, 2) +# define VUINT8x2_MAX_DEFINED +#endif +#ifndef VUINT8x2_RSHIFT_DEFINED +VEC_GENERIC_RSHIFT(u, 8, 2) +# define VUINT8x2_RSHIFT_DEFINED +#endif +#ifndef VUINT8x2_LRSHIFT_DEFINED +VEC_GENERIC_LRSHIFT(u, 8, 2) +# define VUINT8x2_LRSHIFT_DEFINED +#endif +#ifndef VUINT8x2_LSHIFT_DEFINED +VEC_GENERIC_LSHIFT(u, 8, 2) +# define VUINT8x2_LSHIFT_DEFINED +#endif + + +/* vuint8x4 */ + +#ifndef VINT8x4_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(/* nothing */, 8, 4, 2) +# define VINT8x4_SPLAT_DEFINED +#endif + +#ifndef VINT8x4_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(/* nothing */, 8, 4, 2) +# define VINT8x4_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VINT8x4_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(/* nothing */, 8, 4, 2) +# define VINT8x4_LOAD_DEFINED +#endif + +#ifndef VINT8x4_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(/* nothing */, 8, 4, 2) +# define VINT8x4_STORE_ALIGNED_DEFINED +#endif + +#ifndef VINT8x4_STORE_DEFINED +VEC_GENERIC_DBL_STORE(/* nothing */, 8, 4, 2) +# define VINT8x4_STORE_DEFINED +#endif + +#ifndef VINT8x4_ADD_DEFINED +VEC_GENERIC_DBL_ADD(/* nothing */, 8, 4, 2) +# define VINT8x4_ADD_DEFINED +#endif + +#ifndef VINT8x4_SUB_DEFINED +VEC_GENERIC_DBL_SUB(/* nothing */, 8, 4, 2) +# define VINT8x4_SUB_DEFINED +#endif + +#ifndef VINT8x4_MUL_DEFINED +VEC_GENERIC_DBL_MUL(/* nothing */, 8, 4, 2) +# define VINT8x4_MUL_DEFINED +#endif + +#ifndef VINT8x4_DIV_DEFINED +VEC_GENERIC_DBL_DIV(/* nothing */, 8, 4, 2) +# define VINT8x4_DIV_DEFINED +#endif + +#ifndef VINT8x4_AVG_DEFINED +VEC_GENERIC_DBL_AVG(/* nothing */, 8, 4, 2) +# define VINT8x4_AVG_DEFINED +#endif + +#ifndef VINT8x4_AND_DEFINED +VEC_GENERIC_DBL_AND(/* nothing */, 8, 4, 2) +# define VINT8x4_AND_DEFINED +#endif + +#ifndef VINT8x4_OR_DEFINED +VEC_GENERIC_DBL_OR(/* nothing */, 8, 4, 2) +# define VINT8x4_OR_DEFINED +#endif + +#ifndef VINT8x4_XOR_DEFINED +VEC_GENERIC_DBL_XOR(/* nothing */, 8, 4, 2) +# define VINT8x4_XOR_DEFINED +#endif + +#ifndef VINT8x4_NOT_DEFINED +VEC_GENERIC_DBL_NOT(/* nothing */, 8, 4, 2) +# define VINT8x4_NOT_DEFINED +#endif + +#ifndef VINT8x4_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(/* nothing */, 8, 4, 2) +# define VINT8x4_CMPLT_DEFINED +#endif + +#ifndef VINT8x4_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(/* nothing */, 8, 4, 2) +# define VINT8x4_CMPEQ_DEFINED +#endif + +#ifndef VINT8x4_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(/* nothing */, 8, 4, 2) +# define VINT8x4_CMPGT_DEFINED +#endif + +#ifndef VINT8x4_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(/* nothing */, 8, 4, 2) +# define VINT8x4_CMPLE_DEFINED +#endif + +#ifndef VINT8x4_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(/* nothing */, 8, 4, 2) +# define VINT8x4_CMPGE_DEFINED +#endif + +#ifndef VINT8x4_MIN_DEFINED +VEC_GENERIC_DBL_MIN(/* nothing */, 8, 4, 2) +# define VINT8x4_MIN_DEFINED +#endif + +#ifndef VINT8x4_MAX_DEFINED +VEC_GENERIC_DBL_MAX(/* nothing */, 8, 4, 2) +# define VINT8x4_MAX_DEFINED +#endif + +#ifndef VINT8x4_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(/* nothing */, 8, 4, 2) +# define VINT8x4_RSHIFT_DEFINED +#endif + +#ifndef VINT8x4_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(/* nothing */, 8, 4, 2) +# define VINT8x4_LRSHIFT_DEFINED +#endif + +#ifndef VINT8x4_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(/* nothing */, 8, 4, 2) +# define VINT8x4_LSHIFT_DEFINED +#endif + + + +/* vint8x4 */ + +#ifndef VUINT8x4_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(u, 8, 4, 2) +# define VUINT8x4_SPLAT_DEFINED +#endif + +#ifndef VUINT8x4_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(u, 8, 4, 2) +# define VUINT8x4_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VUINT8x4_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(u, 8, 4, 2) +# define VUINT8x4_LOAD_DEFINED +#endif + +#ifndef VUINT8x4_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(u, 8, 4, 2) +# define VUINT8x4_STORE_ALIGNED_DEFINED +#endif + +#ifndef VUINT8x4_STORE_DEFINED +VEC_GENERIC_DBL_STORE(u, 8, 4, 2) +# define VUINT8x4_STORE_DEFINED +#endif + +#ifndef VUINT8x4_ADD_DEFINED +VEC_GENERIC_DBL_ADD(u, 8, 4, 2) +# define VUINT8x4_ADD_DEFINED +#endif + +#ifndef VUINT8x4_SUB_DEFINED +VEC_GENERIC_DBL_SUB(u, 8, 4, 2) +# define VUINT8x4_SUB_DEFINED +#endif + +#ifndef VUINT8x4_MUL_DEFINED +VEC_GENERIC_DBL_MUL(u, 8, 4, 2) +# define VUINT8x4_MUL_DEFINED +#endif + +#ifndef VUINT8x4_DIV_DEFINED +VEC_GENERIC_DBL_DIV(u, 8, 4, 2) +# define VUINT8x4_DIV_DEFINED +#endif + +#ifndef VUINT8x4_AVG_DEFINED +VEC_GENERIC_DBL_AVG(u, 8, 4, 2) +# define VUINT8x4_AVG_DEFINED +#endif + +#ifndef VUINT8x4_AND_DEFINED +VEC_GENERIC_DBL_AND(u, 8, 4, 2) +# define VUINT8x4_AND_DEFINED +#endif + +#ifndef VUINT8x4_OR_DEFINED +VEC_GENERIC_DBL_OR(u, 8, 4, 2) +# define VUINT8x4_OR_DEFINED +#endif + +#ifndef VUINT8x4_XOR_DEFINED +VEC_GENERIC_DBL_XOR(u, 8, 4, 2) +# define VUINT8x4_XOR_DEFINED +#endif + +#ifndef VUINT8x4_NOT_DEFINED +VEC_GENERIC_DBL_NOT(u, 8, 4, 2) +# define VUINT8x4_NOT_DEFINED +#endif + +#ifndef VUINT8x4_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(u, 8, 4, 2) +# define VUINT8x4_CMPLT_DEFINED +#endif + +#ifndef VUINT8x4_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(u, 8, 4, 2) +# define VUINT8x4_CMPEQ_DEFINED +#endif + +#ifndef VUINT8x4_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(u, 8, 4, 2) +# define VUINT8x4_CMPGT_DEFINED +#endif + +#ifndef VUINT8x4_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(u, 8, 4, 2) +# define VUINT8x4_CMPLE_DEFINED +#endif + +#ifndef VUINT8x4_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(u, 8, 4, 2) +# define VUINT8x4_CMPGE_DEFINED +#endif + +#ifndef VUINT8x4_MIN_DEFINED +VEC_GENERIC_DBL_MIN(u, 8, 4, 2) +# define VUINT8x4_MIN_DEFINED +#endif + +#ifndef VUINT8x4_MAX_DEFINED +VEC_GENERIC_DBL_MAX(u, 8, 4, 2) +# define VUINT8x4_MAX_DEFINED +#endif + +#ifndef VUINT8x4_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(u, 8, 4, 2) +# define VUINT8x4_RSHIFT_DEFINED +#endif + +#ifndef VUINT8x4_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(u, 8, 4, 2) +# define VUINT8x4_LRSHIFT_DEFINED +#endif + +#ifndef VUINT8x4_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(u, 8, 4, 2) +# define VUINT8x4_LSHIFT_DEFINED +#endif + + + +/* vuint8x8 */ + +#ifndef VINT8x8_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(/* nothing */, 8, 8, 4) +# define VINT8x8_SPLAT_DEFINED +#endif + +#ifndef VINT8x8_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(/* nothing */, 8, 8, 4) +# define VINT8x8_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VINT8x8_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(/* nothing */, 8, 8, 4) +# define VINT8x8_LOAD_DEFINED +#endif + +#ifndef VINT8x8_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(/* nothing */, 8, 8, 4) +# define VINT8x8_STORE_ALIGNED_DEFINED +#endif + +#ifndef VINT8x8_STORE_DEFINED +VEC_GENERIC_DBL_STORE(/* nothing */, 8, 8, 4) +# define VINT8x8_STORE_DEFINED +#endif + +#ifndef VINT8x8_ADD_DEFINED +VEC_GENERIC_DBL_ADD(/* nothing */, 8, 8, 4) +# define VINT8x8_ADD_DEFINED +#endif + +#ifndef VINT8x8_SUB_DEFINED +VEC_GENERIC_DBL_SUB(/* nothing */, 8, 8, 4) +# define VINT8x8_SUB_DEFINED +#endif + +#ifndef VINT8x8_MUL_DEFINED +VEC_GENERIC_DBL_MUL(/* nothing */, 8, 8, 4) +# define VINT8x8_MUL_DEFINED +#endif + +#ifndef VINT8x8_DIV_DEFINED +VEC_GENERIC_DBL_DIV(/* nothing */, 8, 8, 4) +# define VINT8x8_DIV_DEFINED +#endif + +#ifndef VINT8x8_AVG_DEFINED +VEC_GENERIC_DBL_AVG(/* nothing */, 8, 8, 4) +# define VINT8x8_AVG_DEFINED +#endif + +#ifndef VINT8x8_AND_DEFINED +VEC_GENERIC_DBL_AND(/* nothing */, 8, 8, 4) +# define VINT8x8_AND_DEFINED +#endif + +#ifndef VINT8x8_OR_DEFINED +VEC_GENERIC_DBL_OR(/* nothing */, 8, 8, 4) +# define VINT8x8_OR_DEFINED +#endif + +#ifndef VINT8x8_XOR_DEFINED +VEC_GENERIC_DBL_XOR(/* nothing */, 8, 8, 4) +# define VINT8x8_XOR_DEFINED +#endif + +#ifndef VINT8x8_NOT_DEFINED +VEC_GENERIC_DBL_NOT(/* nothing */, 8, 8, 4) +# define VINT8x8_NOT_DEFINED +#endif + +#ifndef VINT8x8_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(/* nothing */, 8, 8, 4) +# define VINT8x8_CMPLT_DEFINED +#endif + +#ifndef VINT8x8_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(/* nothing */, 8, 8, 4) +# define VINT8x8_CMPEQ_DEFINED +#endif + +#ifndef VINT8x8_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(/* nothing */, 8, 8, 4) +# define VINT8x8_CMPGT_DEFINED +#endif + +#ifndef VINT8x8_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(/* nothing */, 8, 8, 4) +# define VINT8x8_CMPLE_DEFINED +#endif + +#ifndef VINT8x8_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(/* nothing */, 8, 8, 4) +# define VINT8x8_CMPGE_DEFINED +#endif + +#ifndef VINT8x8_MIN_DEFINED +VEC_GENERIC_DBL_MIN(/* nothing */, 8, 8, 4) +# define VINT8x8_MIN_DEFINED +#endif + +#ifndef VINT8x8_MAX_DEFINED +VEC_GENERIC_DBL_MAX(/* nothing */, 8, 8, 4) +# define VINT8x8_MAX_DEFINED +#endif + +#ifndef VINT8x8_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(/* nothing */, 8, 8, 4) +# define VINT8x8_RSHIFT_DEFINED +#endif + +#ifndef VINT8x8_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(/* nothing */, 8, 8, 4) +# define VINT8x8_LRSHIFT_DEFINED +#endif + +#ifndef VINT8x8_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(/* nothing */, 8, 8, 4) +# define VINT8x8_LSHIFT_DEFINED +#endif + + + +/* vint8x8 */ + +#ifndef VUINT8x8_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(u, 8, 8, 4) +# define VUINT8x8_SPLAT_DEFINED +#endif + +#ifndef VUINT8x8_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(u, 8, 8, 4) +# define VUINT8x8_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VUINT8x8_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(u, 8, 8, 4) +# define VUINT8x8_LOAD_DEFINED +#endif + +#ifndef VUINT8x8_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(u, 8, 8, 4) +# define VUINT8x8_STORE_ALIGNED_DEFINED +#endif + +#ifndef VUINT8x8_STORE_DEFINED +VEC_GENERIC_DBL_STORE(u, 8, 8, 4) +# define VUINT8x8_STORE_DEFINED +#endif + +#ifndef VUINT8x8_ADD_DEFINED +VEC_GENERIC_DBL_ADD(u, 8, 8, 4) +# define VUINT8x8_ADD_DEFINED +#endif + +#ifndef VUINT8x8_SUB_DEFINED +VEC_GENERIC_DBL_SUB(u, 8, 8, 4) +# define VUINT8x8_SUB_DEFINED +#endif + +#ifndef VUINT8x8_MUL_DEFINED +VEC_GENERIC_DBL_MUL(u, 8, 8, 4) +# define VUINT8x8_MUL_DEFINED +#endif + +#ifndef VUINT8x8_DIV_DEFINED +VEC_GENERIC_DBL_DIV(u, 8, 8, 4) +# define VUINT8x8_DIV_DEFINED +#endif + +#ifndef VUINT8x8_AVG_DEFINED +VEC_GENERIC_DBL_AVG(u, 8, 8, 4) +# define VUINT8x8_AVG_DEFINED +#endif + +#ifndef VUINT8x8_AND_DEFINED +VEC_GENERIC_DBL_AND(u, 8, 8, 4) +# define VUINT8x8_AND_DEFINED +#endif + +#ifndef VUINT8x8_OR_DEFINED +VEC_GENERIC_DBL_OR(u, 8, 8, 4) +# define VUINT8x8_OR_DEFINED +#endif + +#ifndef VUINT8x8_XOR_DEFINED +VEC_GENERIC_DBL_XOR(u, 8, 8, 4) +# define VUINT8x8_XOR_DEFINED +#endif + +#ifndef VUINT8x8_NOT_DEFINED +VEC_GENERIC_DBL_NOT(u, 8, 8, 4) +# define VUINT8x8_NOT_DEFINED +#endif + +#ifndef VUINT8x8_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(u, 8, 8, 4) +# define VUINT8x8_CMPLT_DEFINED +#endif + +#ifndef VUINT8x8_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(u, 8, 8, 4) +# define VUINT8x8_CMPEQ_DEFINED +#endif + +#ifndef VUINT8x8_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(u, 8, 8, 4) +# define VUINT8x8_CMPGT_DEFINED +#endif + +#ifndef VUINT8x8_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(u, 8, 8, 4) +# define VUINT8x8_CMPLE_DEFINED +#endif + +#ifndef VUINT8x8_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(u, 8, 8, 4) +# define VUINT8x8_CMPGE_DEFINED +#endif + +#ifndef VUINT8x8_MIN_DEFINED +VEC_GENERIC_DBL_MIN(u, 8, 8, 4) +# define VUINT8x8_MIN_DEFINED +#endif + +#ifndef VUINT8x8_MAX_DEFINED +VEC_GENERIC_DBL_MAX(u, 8, 8, 4) +# define VUINT8x8_MAX_DEFINED +#endif + +#ifndef VUINT8x8_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(u, 8, 8, 4) +# define VUINT8x8_RSHIFT_DEFINED +#endif + +#ifndef VUINT8x8_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(u, 8, 8, 4) +# define VUINT8x8_LRSHIFT_DEFINED +#endif + +#ifndef VUINT8x8_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(u, 8, 8, 4) +# define VUINT8x8_LSHIFT_DEFINED +#endif + + + +/* vuint8x16 */ + +#ifndef VINT8x16_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(/* nothing */, 8, 16, 8) +# define VINT8x16_SPLAT_DEFINED +#endif + +#ifndef VINT8x16_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(/* nothing */, 8, 16, 8) +# define VINT8x16_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VINT8x16_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(/* nothing */, 8, 16, 8) +# define VINT8x16_LOAD_DEFINED +#endif + +#ifndef VINT8x16_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(/* nothing */, 8, 16, 8) +# define VINT8x16_STORE_ALIGNED_DEFINED +#endif + +#ifndef VINT8x16_STORE_DEFINED +VEC_GENERIC_DBL_STORE(/* nothing */, 8, 16, 8) +# define VINT8x16_STORE_DEFINED +#endif + +#ifndef VINT8x16_ADD_DEFINED +VEC_GENERIC_DBL_ADD(/* nothing */, 8, 16, 8) +# define VINT8x16_ADD_DEFINED +#endif + +#ifndef VINT8x16_SUB_DEFINED +VEC_GENERIC_DBL_SUB(/* nothing */, 8, 16, 8) +# define VINT8x16_SUB_DEFINED +#endif + +#ifndef VINT8x16_MUL_DEFINED +VEC_GENERIC_DBL_MUL(/* nothing */, 8, 16, 8) +# define VINT8x16_MUL_DEFINED +#endif + +#ifndef VINT8x16_DIV_DEFINED +VEC_GENERIC_DBL_DIV(/* nothing */, 8, 16, 8) +# define VINT8x16_DIV_DEFINED +#endif + +#ifndef VINT8x16_AVG_DEFINED +VEC_GENERIC_DBL_AVG(/* nothing */, 8, 16, 8) +# define VINT8x16_AVG_DEFINED +#endif + +#ifndef VINT8x16_AND_DEFINED +VEC_GENERIC_DBL_AND(/* nothing */, 8, 16, 8) +# define VINT8x16_AND_DEFINED +#endif + +#ifndef VINT8x16_OR_DEFINED +VEC_GENERIC_DBL_OR(/* nothing */, 8, 16, 8) +# define VINT8x16_OR_DEFINED +#endif + +#ifndef VINT8x16_XOR_DEFINED +VEC_GENERIC_DBL_XOR(/* nothing */, 8, 16, 8) +# define VINT8x16_XOR_DEFINED +#endif + +#ifndef VINT8x16_NOT_DEFINED +VEC_GENERIC_DBL_NOT(/* nothing */, 8, 16, 8) +# define VINT8x16_NOT_DEFINED +#endif + +#ifndef VINT8x16_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(/* nothing */, 8, 16, 8) +# define VINT8x16_CMPLT_DEFINED +#endif + +#ifndef VINT8x16_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(/* nothing */, 8, 16, 8) +# define VINT8x16_CMPEQ_DEFINED +#endif + +#ifndef VINT8x16_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(/* nothing */, 8, 16, 8) +# define VINT8x16_CMPGT_DEFINED +#endif + +#ifndef VINT8x16_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(/* nothing */, 8, 16, 8) +# define VINT8x16_CMPLE_DEFINED +#endif + +#ifndef VINT8x16_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(/* nothing */, 8, 16, 8) +# define VINT8x16_CMPGE_DEFINED +#endif + +#ifndef VINT8x16_MIN_DEFINED +VEC_GENERIC_DBL_MIN(/* nothing */, 8, 16, 8) +# define VINT8x16_MIN_DEFINED +#endif + +#ifndef VINT8x16_MAX_DEFINED +VEC_GENERIC_DBL_MAX(/* nothing */, 8, 16, 8) +# define VINT8x16_MAX_DEFINED +#endif + +#ifndef VINT8x16_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(/* nothing */, 8, 16, 8) +# define VINT8x16_RSHIFT_DEFINED +#endif + +#ifndef VINT8x16_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(/* nothing */, 8, 16, 8) +# define VINT8x16_LRSHIFT_DEFINED +#endif + +#ifndef VINT8x16_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(/* nothing */, 8, 16, 8) +# define VINT8x16_LSHIFT_DEFINED +#endif + + + +/* vint8x16 */ + +#ifndef VUINT8x16_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(u, 8, 16, 8) +# define VUINT8x16_SPLAT_DEFINED +#endif + +#ifndef VUINT8x16_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(u, 8, 16, 8) +# define VUINT8x16_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VUINT8x16_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(u, 8, 16, 8) +# define VUINT8x16_LOAD_DEFINED +#endif + +#ifndef VUINT8x16_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(u, 8, 16, 8) +# define VUINT8x16_STORE_ALIGNED_DEFINED +#endif + +#ifndef VUINT8x16_STORE_DEFINED +VEC_GENERIC_DBL_STORE(u, 8, 16, 8) +# define VUINT8x16_STORE_DEFINED +#endif + +#ifndef VUINT8x16_ADD_DEFINED +VEC_GENERIC_DBL_ADD(u, 8, 16, 8) +# define VUINT8x16_ADD_DEFINED +#endif + +#ifndef VUINT8x16_SUB_DEFINED +VEC_GENERIC_DBL_SUB(u, 8, 16, 8) +# define VUINT8x16_SUB_DEFINED +#endif + +#ifndef VUINT8x16_MUL_DEFINED +VEC_GENERIC_DBL_MUL(u, 8, 16, 8) +# define VUINT8x16_MUL_DEFINED +#endif + +#ifndef VUINT8x16_DIV_DEFINED +VEC_GENERIC_DBL_DIV(u, 8, 16, 8) +# define VUINT8x16_DIV_DEFINED +#endif + +#ifndef VUINT8x16_AVG_DEFINED +VEC_GENERIC_DBL_AVG(u, 8, 16, 8) +# define VUINT8x16_AVG_DEFINED +#endif + +#ifndef VUINT8x16_AND_DEFINED +VEC_GENERIC_DBL_AND(u, 8, 16, 8) +# define VUINT8x16_AND_DEFINED +#endif + +#ifndef VUINT8x16_OR_DEFINED +VEC_GENERIC_DBL_OR(u, 8, 16, 8) +# define VUINT8x16_OR_DEFINED +#endif + +#ifndef VUINT8x16_XOR_DEFINED +VEC_GENERIC_DBL_XOR(u, 8, 16, 8) +# define VUINT8x16_XOR_DEFINED +#endif + +#ifndef VUINT8x16_NOT_DEFINED +VEC_GENERIC_DBL_NOT(u, 8, 16, 8) +# define VUINT8x16_NOT_DEFINED +#endif + +#ifndef VUINT8x16_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(u, 8, 16, 8) +# define VUINT8x16_CMPLT_DEFINED +#endif + +#ifndef VUINT8x16_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(u, 8, 16, 8) +# define VUINT8x16_CMPEQ_DEFINED +#endif + +#ifndef VUINT8x16_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(u, 8, 16, 8) +# define VUINT8x16_CMPGT_DEFINED +#endif + +#ifndef VUINT8x16_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(u, 8, 16, 8) +# define VUINT8x16_CMPLE_DEFINED +#endif + +#ifndef VUINT8x16_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(u, 8, 16, 8) +# define VUINT8x16_CMPGE_DEFINED +#endif + +#ifndef VUINT8x16_MIN_DEFINED +VEC_GENERIC_DBL_MIN(u, 8, 16, 8) +# define VUINT8x16_MIN_DEFINED +#endif + +#ifndef VUINT8x16_MAX_DEFINED +VEC_GENERIC_DBL_MAX(u, 8, 16, 8) +# define VUINT8x16_MAX_DEFINED +#endif + +#ifndef VUINT8x16_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(u, 8, 16, 8) +# define VUINT8x16_RSHIFT_DEFINED +#endif + +#ifndef VUINT8x16_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(u, 8, 16, 8) +# define VUINT8x16_LRSHIFT_DEFINED +#endif + +#ifndef VUINT8x16_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(u, 8, 16, 8) +# define VUINT8x16_LSHIFT_DEFINED +#endif + + + +/* vuint8x32 */ + +#ifndef VINT8x32_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(/* nothing */, 8, 32, 16) +# define VINT8x32_SPLAT_DEFINED +#endif + +#ifndef VINT8x32_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(/* nothing */, 8, 32, 16) +# define VINT8x32_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VINT8x32_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(/* nothing */, 8, 32, 16) +# define VINT8x32_LOAD_DEFINED +#endif + +#ifndef VINT8x32_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(/* nothing */, 8, 32, 16) +# define VINT8x32_STORE_ALIGNED_DEFINED +#endif + +#ifndef VINT8x32_STORE_DEFINED +VEC_GENERIC_DBL_STORE(/* nothing */, 8, 32, 16) +# define VINT8x32_STORE_DEFINED +#endif + +#ifndef VINT8x32_ADD_DEFINED +VEC_GENERIC_DBL_ADD(/* nothing */, 8, 32, 16) +# define VINT8x32_ADD_DEFINED +#endif + +#ifndef VINT8x32_SUB_DEFINED +VEC_GENERIC_DBL_SUB(/* nothing */, 8, 32, 16) +# define VINT8x32_SUB_DEFINED +#endif + +#ifndef VINT8x32_MUL_DEFINED +VEC_GENERIC_DBL_MUL(/* nothing */, 8, 32, 16) +# define VINT8x32_MUL_DEFINED +#endif + +#ifndef VINT8x32_DIV_DEFINED +VEC_GENERIC_DBL_DIV(/* nothing */, 8, 32, 16) +# define VINT8x32_DIV_DEFINED +#endif + +#ifndef VINT8x32_AVG_DEFINED +VEC_GENERIC_DBL_AVG(/* nothing */, 8, 32, 16) +# define VINT8x32_AVG_DEFINED +#endif + +#ifndef VINT8x32_AND_DEFINED +VEC_GENERIC_DBL_AND(/* nothing */, 8, 32, 16) +# define VINT8x32_AND_DEFINED +#endif + +#ifndef VINT8x32_OR_DEFINED +VEC_GENERIC_DBL_OR(/* nothing */, 8, 32, 16) +# define VINT8x32_OR_DEFINED +#endif + +#ifndef VINT8x32_XOR_DEFINED +VEC_GENERIC_DBL_XOR(/* nothing */, 8, 32, 16) +# define VINT8x32_XOR_DEFINED +#endif + +#ifndef VINT8x32_NOT_DEFINED +VEC_GENERIC_DBL_NOT(/* nothing */, 8, 32, 16) +# define VINT8x32_NOT_DEFINED +#endif + +#ifndef VINT8x32_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(/* nothing */, 8, 32, 16) +# define VINT8x32_CMPLT_DEFINED +#endif + +#ifndef VINT8x32_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(/* nothing */, 8, 32, 16) +# define VINT8x32_CMPEQ_DEFINED +#endif + +#ifndef VINT8x32_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(/* nothing */, 8, 32, 16) +# define VINT8x32_CMPGT_DEFINED +#endif + +#ifndef VINT8x32_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(/* nothing */, 8, 32, 16) +# define VINT8x32_CMPLE_DEFINED +#endif + +#ifndef VINT8x32_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(/* nothing */, 8, 32, 16) +# define VINT8x32_CMPGE_DEFINED +#endif + +#ifndef VINT8x32_MIN_DEFINED +VEC_GENERIC_DBL_MIN(/* nothing */, 8, 32, 16) +# define VINT8x32_MIN_DEFINED +#endif + +#ifndef VINT8x32_MAX_DEFINED +VEC_GENERIC_DBL_MAX(/* nothing */, 8, 32, 16) +# define VINT8x32_MAX_DEFINED +#endif + +#ifndef VINT8x32_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(/* nothing */, 8, 32, 16) +# define VINT8x32_RSHIFT_DEFINED +#endif + +#ifndef VINT8x32_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(/* nothing */, 8, 32, 16) +# define VINT8x32_LRSHIFT_DEFINED +#endif + +#ifndef VINT8x32_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(/* nothing */, 8, 32, 16) +# define VINT8x32_LSHIFT_DEFINED +#endif + + + +/* vint8x32 */ + +#ifndef VUINT8x32_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(u, 8, 32, 16) +# define VUINT8x32_SPLAT_DEFINED +#endif + +#ifndef VUINT8x32_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(u, 8, 32, 16) +# define VUINT8x32_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VUINT8x32_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(u, 8, 32, 16) +# define VUINT8x32_LOAD_DEFINED +#endif + +#ifndef VUINT8x32_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(u, 8, 32, 16) +# define VUINT8x32_STORE_ALIGNED_DEFINED +#endif + +#ifndef VUINT8x32_STORE_DEFINED +VEC_GENERIC_DBL_STORE(u, 8, 32, 16) +# define VUINT8x32_STORE_DEFINED +#endif + +#ifndef VUINT8x32_ADD_DEFINED +VEC_GENERIC_DBL_ADD(u, 8, 32, 16) +# define VUINT8x32_ADD_DEFINED +#endif + +#ifndef VUINT8x32_SUB_DEFINED +VEC_GENERIC_DBL_SUB(u, 8, 32, 16) +# define VUINT8x32_SUB_DEFINED +#endif + +#ifndef VUINT8x32_MUL_DEFINED +VEC_GENERIC_DBL_MUL(u, 8, 32, 16) +# define VUINT8x32_MUL_DEFINED +#endif + +#ifndef VUINT8x32_DIV_DEFINED +VEC_GENERIC_DBL_DIV(u, 8, 32, 16) +# define VUINT8x32_DIV_DEFINED +#endif + +#ifndef VUINT8x32_AVG_DEFINED +VEC_GENERIC_DBL_AVG(u, 8, 32, 16) +# define VUINT8x32_AVG_DEFINED +#endif + +#ifndef VUINT8x32_AND_DEFINED +VEC_GENERIC_DBL_AND(u, 8, 32, 16) +# define VUINT8x32_AND_DEFINED +#endif + +#ifndef VUINT8x32_OR_DEFINED +VEC_GENERIC_DBL_OR(u, 8, 32, 16) +# define VUINT8x32_OR_DEFINED +#endif + +#ifndef VUINT8x32_XOR_DEFINED +VEC_GENERIC_DBL_XOR(u, 8, 32, 16) +# define VUINT8x32_XOR_DEFINED +#endif + +#ifndef VUINT8x32_NOT_DEFINED +VEC_GENERIC_DBL_NOT(u, 8, 32, 16) +# define VUINT8x32_NOT_DEFINED +#endif + +#ifndef VUINT8x32_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(u, 8, 32, 16) +# define VUINT8x32_CMPLT_DEFINED +#endif + +#ifndef VUINT8x32_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(u, 8, 32, 16) +# define VUINT8x32_CMPEQ_DEFINED +#endif + +#ifndef VUINT8x32_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(u, 8, 32, 16) +# define VUINT8x32_CMPGT_DEFINED +#endif + +#ifndef VUINT8x32_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(u, 8, 32, 16) +# define VUINT8x32_CMPLE_DEFINED +#endif + +#ifndef VUINT8x32_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(u, 8, 32, 16) +# define VUINT8x32_CMPGE_DEFINED +#endif + +#ifndef VUINT8x32_MIN_DEFINED +VEC_GENERIC_DBL_MIN(u, 8, 32, 16) +# define VUINT8x32_MIN_DEFINED +#endif + +#ifndef VUINT8x32_MAX_DEFINED +VEC_GENERIC_DBL_MAX(u, 8, 32, 16) +# define VUINT8x32_MAX_DEFINED +#endif + +#ifndef VUINT8x32_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(u, 8, 32, 16) +# define VUINT8x32_RSHIFT_DEFINED +#endif + +#ifndef VUINT8x32_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(u, 8, 32, 16) +# define VUINT8x32_LRSHIFT_DEFINED +#endif + +#ifndef VUINT8x32_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(u, 8, 32, 16) +# define VUINT8x32_LSHIFT_DEFINED +#endif + + + +/* vuint8x64 */ + +#ifndef VINT8x64_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(/* nothing */, 8, 64, 32) +# define VINT8x64_SPLAT_DEFINED +#endif + +#ifndef VINT8x64_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(/* nothing */, 8, 64, 32) +# define VINT8x64_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VINT8x64_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(/* nothing */, 8, 64, 32) +# define VINT8x64_LOAD_DEFINED +#endif + +#ifndef VINT8x64_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(/* nothing */, 8, 64, 32) +# define VINT8x64_STORE_ALIGNED_DEFINED +#endif + +#ifndef VINT8x64_STORE_DEFINED +VEC_GENERIC_DBL_STORE(/* nothing */, 8, 64, 32) +# define VINT8x64_STORE_DEFINED +#endif + +#ifndef VINT8x64_ADD_DEFINED +VEC_GENERIC_DBL_ADD(/* nothing */, 8, 64, 32) +# define VINT8x64_ADD_DEFINED +#endif + +#ifndef VINT8x64_SUB_DEFINED +VEC_GENERIC_DBL_SUB(/* nothing */, 8, 64, 32) +# define VINT8x64_SUB_DEFINED +#endif + +#ifndef VINT8x64_MUL_DEFINED +VEC_GENERIC_DBL_MUL(/* nothing */, 8, 64, 32) +# define VINT8x64_MUL_DEFINED +#endif + +#ifndef VINT8x64_DIV_DEFINED +VEC_GENERIC_DBL_DIV(/* nothing */, 8, 64, 32) +# define VINT8x64_DIV_DEFINED +#endif + +#ifndef VINT8x64_AVG_DEFINED +VEC_GENERIC_DBL_AVG(/* nothing */, 8, 64, 32) +# define VINT8x64_AVG_DEFINED +#endif + +#ifndef VINT8x64_AND_DEFINED +VEC_GENERIC_DBL_AND(/* nothing */, 8, 64, 32) +# define VINT8x64_AND_DEFINED +#endif + +#ifndef VINT8x64_OR_DEFINED +VEC_GENERIC_DBL_OR(/* nothing */, 8, 64, 32) +# define VINT8x64_OR_DEFINED +#endif + +#ifndef VINT8x64_XOR_DEFINED +VEC_GENERIC_DBL_XOR(/* nothing */, 8, 64, 32) +# define VINT8x64_XOR_DEFINED +#endif + +#ifndef VINT8x64_NOT_DEFINED +VEC_GENERIC_DBL_NOT(/* nothing */, 8, 64, 32) +# define VINT8x64_NOT_DEFINED +#endif + +#ifndef VINT8x64_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(/* nothing */, 8, 64, 32) +# define VINT8x64_CMPLT_DEFINED +#endif + +#ifndef VINT8x64_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(/* nothing */, 8, 64, 32) +# define VINT8x64_CMPEQ_DEFINED +#endif + +#ifndef VINT8x64_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(/* nothing */, 8, 64, 32) +# define VINT8x64_CMPGT_DEFINED +#endif + +#ifndef VINT8x64_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(/* nothing */, 8, 64, 32) +# define VINT8x64_CMPLE_DEFINED +#endif + +#ifndef VINT8x64_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(/* nothing */, 8, 64, 32) +# define VINT8x64_CMPGE_DEFINED +#endif + +#ifndef VINT8x64_MIN_DEFINED +VEC_GENERIC_DBL_MIN(/* nothing */, 8, 64, 32) +# define VINT8x64_MIN_DEFINED +#endif + +#ifndef VINT8x64_MAX_DEFINED +VEC_GENERIC_DBL_MAX(/* nothing */, 8, 64, 32) +# define VINT8x64_MAX_DEFINED +#endif + +#ifndef VINT8x64_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(/* nothing */, 8, 64, 32) +# define VINT8x64_RSHIFT_DEFINED +#endif + +#ifndef VINT8x64_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(/* nothing */, 8, 64, 32) +# define VINT8x64_LRSHIFT_DEFINED +#endif + +#ifndef VINT8x64_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(/* nothing */, 8, 64, 32) +# define VINT8x64_LSHIFT_DEFINED +#endif + + + +/* vint8x64 */ + +#ifndef VUINT8x64_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(u, 8, 64, 32) +# define VUINT8x64_SPLAT_DEFINED +#endif + +#ifndef VUINT8x64_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(u, 8, 64, 32) +# define VUINT8x64_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VUINT8x64_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(u, 8, 64, 32) +# define VUINT8x64_LOAD_DEFINED +#endif + +#ifndef VUINT8x64_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(u, 8, 64, 32) +# define VUINT8x64_STORE_ALIGNED_DEFINED +#endif + +#ifndef VUINT8x64_STORE_DEFINED +VEC_GENERIC_DBL_STORE(u, 8, 64, 32) +# define VUINT8x64_STORE_DEFINED +#endif + +#ifndef VUINT8x64_ADD_DEFINED +VEC_GENERIC_DBL_ADD(u, 8, 64, 32) +# define VUINT8x64_ADD_DEFINED +#endif + +#ifndef VUINT8x64_SUB_DEFINED +VEC_GENERIC_DBL_SUB(u, 8, 64, 32) +# define VUINT8x64_SUB_DEFINED +#endif + +#ifndef VUINT8x64_MUL_DEFINED +VEC_GENERIC_DBL_MUL(u, 8, 64, 32) +# define VUINT8x64_MUL_DEFINED +#endif + +#ifndef VUINT8x64_DIV_DEFINED +VEC_GENERIC_DBL_DIV(u, 8, 64, 32) +# define VUINT8x64_DIV_DEFINED +#endif + +#ifndef VUINT8x64_AVG_DEFINED +VEC_GENERIC_DBL_AVG(u, 8, 64, 32) +# define VUINT8x64_AVG_DEFINED +#endif + +#ifndef VUINT8x64_AND_DEFINED +VEC_GENERIC_DBL_AND(u, 8, 64, 32) +# define VUINT8x64_AND_DEFINED +#endif + +#ifndef VUINT8x64_OR_DEFINED +VEC_GENERIC_DBL_OR(u, 8, 64, 32) +# define VUINT8x64_OR_DEFINED +#endif + +#ifndef VUINT8x64_XOR_DEFINED +VEC_GENERIC_DBL_XOR(u, 8, 64, 32) +# define VUINT8x64_XOR_DEFINED +#endif + +#ifndef VUINT8x64_NOT_DEFINED +VEC_GENERIC_DBL_NOT(u, 8, 64, 32) +# define VUINT8x64_NOT_DEFINED +#endif + +#ifndef VUINT8x64_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(u, 8, 64, 32) +# define VUINT8x64_CMPLT_DEFINED +#endif + +#ifndef VUINT8x64_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(u, 8, 64, 32) +# define VUINT8x64_CMPEQ_DEFINED +#endif + +#ifndef VUINT8x64_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(u, 8, 64, 32) +# define VUINT8x64_CMPGT_DEFINED +#endif + +#ifndef VUINT8x64_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(u, 8, 64, 32) +# define VUINT8x64_CMPLE_DEFINED +#endif + +#ifndef VUINT8x64_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(u, 8, 64, 32) +# define VUINT8x64_CMPGE_DEFINED +#endif + +#ifndef VUINT8x64_MIN_DEFINED +VEC_GENERIC_DBL_MIN(u, 8, 64, 32) +# define VUINT8x64_MIN_DEFINED +#endif + +#ifndef VUINT8x64_MAX_DEFINED +VEC_GENERIC_DBL_MAX(u, 8, 64, 32) +# define VUINT8x64_MAX_DEFINED +#endif + +#ifndef VUINT8x64_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(u, 8, 64, 32) +# define VUINT8x64_RSHIFT_DEFINED +#endif + +#ifndef VUINT8x64_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(u, 8, 64, 32) +# define VUINT8x64_LRSHIFT_DEFINED +#endif + +#ifndef VUINT8x64_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(u, 8, 64, 32) +# define VUINT8x64_LSHIFT_DEFINED +#endif + + + +/* vuint16x2 */ + +#ifndef VINT16x2_SPLAT_DEFINED +VEC_GENERIC_SPLAT(/* nothing */, 16, 2) +# define VINT16x2_SPLAT_DEFINED +#endif +#ifndef VINT16x2_LOAD_ALIGNED_DEFINED +VEC_GENERIC_LOAD_ALIGNED(/* nothing */, 16, 2) +# define VINT16x2_LOAD_ALIGNED_DEFINED +#endif +#ifndef VINT16x2_LOAD_DEFINED +VEC_GENERIC_LOAD(/* nothing */, 16, 2) +# define VINT16x2_LOAD_DEFINED +#endif +#ifndef VINT16x2_STORE_ALIGNED_DEFINED +VEC_GENERIC_STORE_ALIGNED(/* nothing */, 16, 2) +# define VINT16x2_STORE_ALIGNED_DEFINED +#endif +#ifndef VINT16x2_STORE_DEFINED +VEC_GENERIC_STORE(/* nothing */, 16, 2) +# define VINT16x2_STORE_DEFINED +#endif +#ifndef VINT16x2_ADD_DEFINED +VEC_GENERIC_ADD(/* nothing */, 16, 2) +# define VINT16x2_ADD_DEFINED +#endif +#ifndef VINT16x2_SUB_DEFINED +VEC_GENERIC_SUB(/* nothing */, 16, 2) +# define VINT16x2_SUB_DEFINED +#endif +#ifndef VINT16x2_MUL_DEFINED +VEC_GENERIC_MUL(/* nothing */, 16, 2) +# define VINT16x2_MUL_DEFINED +#endif +#ifndef VINT16x2_DIV_DEFINED +VEC_GENERIC_DIV(/* nothing */, 16, 2) +# define VINT16x2_DIV_DEFINED +#endif +#ifndef VINT16x2_AVG_DEFINED +VEC_GENERIC_AVG(/* nothing */, 16, 2) +# define VINT16x2_AVG_DEFINED +#endif +#ifndef VINT16x2_AND_DEFINED +VEC_GENERIC_AND(/* nothing */, 16, 2) +# define VINT16x2_AND_DEFINED +#endif +#ifndef VINT16x2_OR_DEFINED +VEC_GENERIC_OR(/* nothing */, 16, 2) +# define VINT16x2_OR_DEFINED +#endif +#ifndef VINT16x2_XOR_DEFINED +VEC_GENERIC_XOR(/* nothing */, 16, 2) +# define VINT16x2_XOR_DEFINED +#endif +#ifndef VINT16x2_NOT_DEFINED +VEC_GENERIC_NOT(/* nothing */, 16, 2) +# define VINT16x2_NOT_DEFINED +#endif +#ifndef VINT16x2_CMPLT_DEFINED +VEC_GENERIC_CMPLT(/* nothing */, 16, 2) +# define VINT16x2_CMPLT_DEFINED +#endif +#ifndef VINT16x2_CMPEQ_DEFINED +VEC_GENERIC_CMPEQ(/* nothing */, 16, 2) +# define VINT16x2_CMPEQ_DEFINED +#endif +#ifndef VINT16x2_CMPGT_DEFINED +VEC_GENERIC_CMPGT(/* nothing */, 16, 2) +# define VINT16x2_CMPGT_DEFINED +#endif +#ifndef VINT16x2_CMPLE_DEFINED +VEC_GENERIC_CMPLE(/* nothing */, 16, 2) +# define VINT16x2_CMPLE_DEFINED +#endif +#ifndef VINT16x2_CMPGE_DEFINED +VEC_GENERIC_CMPGE(/* nothing */, 16, 2) +# define VINT16x2_CMPGE_DEFINED +#endif +#ifndef VINT16x2_MIN_DEFINED +VEC_GENERIC_MIN(/* nothing */, 16, 2) +# define VINT16x2_MIN_DEFINED +#endif +#ifndef VINT16x2_MAX_DEFINED +VEC_GENERIC_MAX(/* nothing */, 16, 2) +# define VINT16x2_MAX_DEFINED +#endif +#ifndef VINT16x2_RSHIFT_DEFINED +VEC_GENERIC_RSHIFT(/* nothing */, 16, 2) +# define VINT16x2_RSHIFT_DEFINED +#endif +#ifndef VINT16x2_LRSHIFT_DEFINED +VEC_GENERIC_LRSHIFT(/* nothing */, 16, 2) +# define VINT16x2_LRSHIFT_DEFINED +#endif +#ifndef VINT16x2_LSHIFT_DEFINED +VEC_GENERIC_LSHIFT(/* nothing */, 16, 2) +# define VINT16x2_LSHIFT_DEFINED +#endif + + +/* vint16x2 */ + +#ifndef VUINT16x2_SPLAT_DEFINED +VEC_GENERIC_SPLAT(u, 16, 2) +# define VUINT16x2_SPLAT_DEFINED +#endif +#ifndef VUINT16x2_LOAD_ALIGNED_DEFINED +VEC_GENERIC_LOAD_ALIGNED(u, 16, 2) +# define VUINT16x2_LOAD_ALIGNED_DEFINED +#endif +#ifndef VUINT16x2_LOAD_DEFINED +VEC_GENERIC_LOAD(u, 16, 2) +# define VUINT16x2_LOAD_DEFINED +#endif +#ifndef VUINT16x2_STORE_ALIGNED_DEFINED +VEC_GENERIC_STORE_ALIGNED(u, 16, 2) +# define VUINT16x2_STORE_ALIGNED_DEFINED +#endif +#ifndef VUINT16x2_STORE_DEFINED +VEC_GENERIC_STORE(u, 16, 2) +# define VUINT16x2_STORE_DEFINED +#endif +#ifndef VUINT16x2_ADD_DEFINED +VEC_GENERIC_ADD(u, 16, 2) +# define VUINT16x2_ADD_DEFINED +#endif +#ifndef VUINT16x2_SUB_DEFINED +VEC_GENERIC_SUB(u, 16, 2) +# define VUINT16x2_SUB_DEFINED +#endif +#ifndef VUINT16x2_MUL_DEFINED +VEC_GENERIC_MUL(u, 16, 2) +# define VUINT16x2_MUL_DEFINED +#endif +#ifndef VUINT16x2_DIV_DEFINED +VEC_GENERIC_DIV(u, 16, 2) +# define VUINT16x2_DIV_DEFINED +#endif +#ifndef VUINT16x2_AVG_DEFINED +VEC_GENERIC_AVG(u, 16, 2) +# define VUINT16x2_AVG_DEFINED +#endif +#ifndef VUINT16x2_AND_DEFINED +VEC_GENERIC_AND(u, 16, 2) +# define VUINT16x2_AND_DEFINED +#endif +#ifndef VUINT16x2_OR_DEFINED +VEC_GENERIC_OR(u, 16, 2) +# define VUINT16x2_OR_DEFINED +#endif +#ifndef VUINT16x2_XOR_DEFINED +VEC_GENERIC_XOR(u, 16, 2) +# define VUINT16x2_XOR_DEFINED +#endif +#ifndef VUINT16x2_NOT_DEFINED +VEC_GENERIC_NOT(u, 16, 2) +# define VUINT16x2_NOT_DEFINED +#endif +#ifndef VUINT16x2_CMPLT_DEFINED +VEC_GENERIC_CMPLT(u, 16, 2) +# define VUINT16x2_CMPLT_DEFINED +#endif +#ifndef VUINT16x2_CMPEQ_DEFINED +VEC_GENERIC_CMPEQ(u, 16, 2) +# define VUINT16x2_CMPEQ_DEFINED +#endif +#ifndef VUINT16x2_CMPGT_DEFINED +VEC_GENERIC_CMPGT(u, 16, 2) +# define VUINT16x2_CMPGT_DEFINED +#endif +#ifndef VUINT16x2_CMPLE_DEFINED +VEC_GENERIC_CMPLE(u, 16, 2) +# define VUINT16x2_CMPLE_DEFINED +#endif +#ifndef VUINT16x2_CMPGE_DEFINED +VEC_GENERIC_CMPGE(u, 16, 2) +# define VUINT16x2_CMPGE_DEFINED +#endif +#ifndef VUINT16x2_MIN_DEFINED +VEC_GENERIC_MIN(u, 16, 2) +# define VUINT16x2_MIN_DEFINED +#endif +#ifndef VUINT16x2_MAX_DEFINED +VEC_GENERIC_MAX(u, 16, 2) +# define VUINT16x2_MAX_DEFINED +#endif +#ifndef VUINT16x2_RSHIFT_DEFINED +VEC_GENERIC_RSHIFT(u, 16, 2) +# define VUINT16x2_RSHIFT_DEFINED +#endif +#ifndef VUINT16x2_LRSHIFT_DEFINED +VEC_GENERIC_LRSHIFT(u, 16, 2) +# define VUINT16x2_LRSHIFT_DEFINED +#endif +#ifndef VUINT16x2_LSHIFT_DEFINED +VEC_GENERIC_LSHIFT(u, 16, 2) +# define VUINT16x2_LSHIFT_DEFINED +#endif + + +/* vuint16x4 */ + +#ifndef VINT16x4_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(/* nothing */, 16, 4, 2) +# define VINT16x4_SPLAT_DEFINED +#endif + +#ifndef VINT16x4_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(/* nothing */, 16, 4, 2) +# define VINT16x4_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VINT16x4_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(/* nothing */, 16, 4, 2) +# define VINT16x4_LOAD_DEFINED +#endif + +#ifndef VINT16x4_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(/* nothing */, 16, 4, 2) +# define VINT16x4_STORE_ALIGNED_DEFINED +#endif + +#ifndef VINT16x4_STORE_DEFINED +VEC_GENERIC_DBL_STORE(/* nothing */, 16, 4, 2) +# define VINT16x4_STORE_DEFINED +#endif + +#ifndef VINT16x4_ADD_DEFINED +VEC_GENERIC_DBL_ADD(/* nothing */, 16, 4, 2) +# define VINT16x4_ADD_DEFINED +#endif + +#ifndef VINT16x4_SUB_DEFINED +VEC_GENERIC_DBL_SUB(/* nothing */, 16, 4, 2) +# define VINT16x4_SUB_DEFINED +#endif + +#ifndef VINT16x4_MUL_DEFINED +VEC_GENERIC_DBL_MUL(/* nothing */, 16, 4, 2) +# define VINT16x4_MUL_DEFINED +#endif + +#ifndef VINT16x4_DIV_DEFINED +VEC_GENERIC_DBL_DIV(/* nothing */, 16, 4, 2) +# define VINT16x4_DIV_DEFINED +#endif + +#ifndef VINT16x4_AVG_DEFINED +VEC_GENERIC_DBL_AVG(/* nothing */, 16, 4, 2) +# define VINT16x4_AVG_DEFINED +#endif + +#ifndef VINT16x4_AND_DEFINED +VEC_GENERIC_DBL_AND(/* nothing */, 16, 4, 2) +# define VINT16x4_AND_DEFINED +#endif + +#ifndef VINT16x4_OR_DEFINED +VEC_GENERIC_DBL_OR(/* nothing */, 16, 4, 2) +# define VINT16x4_OR_DEFINED +#endif + +#ifndef VINT16x4_XOR_DEFINED +VEC_GENERIC_DBL_XOR(/* nothing */, 16, 4, 2) +# define VINT16x4_XOR_DEFINED +#endif + +#ifndef VINT16x4_NOT_DEFINED +VEC_GENERIC_DBL_NOT(/* nothing */, 16, 4, 2) +# define VINT16x4_NOT_DEFINED +#endif + +#ifndef VINT16x4_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(/* nothing */, 16, 4, 2) +# define VINT16x4_CMPLT_DEFINED +#endif + +#ifndef VINT16x4_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(/* nothing */, 16, 4, 2) +# define VINT16x4_CMPEQ_DEFINED +#endif + +#ifndef VINT16x4_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(/* nothing */, 16, 4, 2) +# define VINT16x4_CMPGT_DEFINED +#endif + +#ifndef VINT16x4_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(/* nothing */, 16, 4, 2) +# define VINT16x4_CMPLE_DEFINED +#endif + +#ifndef VINT16x4_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(/* nothing */, 16, 4, 2) +# define VINT16x4_CMPGE_DEFINED +#endif + +#ifndef VINT16x4_MIN_DEFINED +VEC_GENERIC_DBL_MIN(/* nothing */, 16, 4, 2) +# define VINT16x4_MIN_DEFINED +#endif + +#ifndef VINT16x4_MAX_DEFINED +VEC_GENERIC_DBL_MAX(/* nothing */, 16, 4, 2) +# define VINT16x4_MAX_DEFINED +#endif + +#ifndef VINT16x4_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(/* nothing */, 16, 4, 2) +# define VINT16x4_RSHIFT_DEFINED +#endif + +#ifndef VINT16x4_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(/* nothing */, 16, 4, 2) +# define VINT16x4_LRSHIFT_DEFINED +#endif + +#ifndef VINT16x4_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(/* nothing */, 16, 4, 2) +# define VINT16x4_LSHIFT_DEFINED +#endif + + + +/* vint16x4 */ + +#ifndef VUINT16x4_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(u, 16, 4, 2) +# define VUINT16x4_SPLAT_DEFINED +#endif + +#ifndef VUINT16x4_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(u, 16, 4, 2) +# define VUINT16x4_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VUINT16x4_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(u, 16, 4, 2) +# define VUINT16x4_LOAD_DEFINED +#endif + +#ifndef VUINT16x4_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(u, 16, 4, 2) +# define VUINT16x4_STORE_ALIGNED_DEFINED +#endif + +#ifndef VUINT16x4_STORE_DEFINED +VEC_GENERIC_DBL_STORE(u, 16, 4, 2) +# define VUINT16x4_STORE_DEFINED +#endif + +#ifndef VUINT16x4_ADD_DEFINED +VEC_GENERIC_DBL_ADD(u, 16, 4, 2) +# define VUINT16x4_ADD_DEFINED +#endif + +#ifndef VUINT16x4_SUB_DEFINED +VEC_GENERIC_DBL_SUB(u, 16, 4, 2) +# define VUINT16x4_SUB_DEFINED +#endif + +#ifndef VUINT16x4_MUL_DEFINED +VEC_GENERIC_DBL_MUL(u, 16, 4, 2) +# define VUINT16x4_MUL_DEFINED +#endif + +#ifndef VUINT16x4_DIV_DEFINED +VEC_GENERIC_DBL_DIV(u, 16, 4, 2) +# define VUINT16x4_DIV_DEFINED +#endif + +#ifndef VUINT16x4_AVG_DEFINED +VEC_GENERIC_DBL_AVG(u, 16, 4, 2) +# define VUINT16x4_AVG_DEFINED +#endif + +#ifndef VUINT16x4_AND_DEFINED +VEC_GENERIC_DBL_AND(u, 16, 4, 2) +# define VUINT16x4_AND_DEFINED +#endif + +#ifndef VUINT16x4_OR_DEFINED +VEC_GENERIC_DBL_OR(u, 16, 4, 2) +# define VUINT16x4_OR_DEFINED +#endif + +#ifndef VUINT16x4_XOR_DEFINED +VEC_GENERIC_DBL_XOR(u, 16, 4, 2) +# define VUINT16x4_XOR_DEFINED +#endif + +#ifndef VUINT16x4_NOT_DEFINED +VEC_GENERIC_DBL_NOT(u, 16, 4, 2) +# define VUINT16x4_NOT_DEFINED +#endif + +#ifndef VUINT16x4_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(u, 16, 4, 2) +# define VUINT16x4_CMPLT_DEFINED +#endif + +#ifndef VUINT16x4_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(u, 16, 4, 2) +# define VUINT16x4_CMPEQ_DEFINED +#endif + +#ifndef VUINT16x4_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(u, 16, 4, 2) +# define VUINT16x4_CMPGT_DEFINED +#endif + +#ifndef VUINT16x4_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(u, 16, 4, 2) +# define VUINT16x4_CMPLE_DEFINED +#endif + +#ifndef VUINT16x4_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(u, 16, 4, 2) +# define VUINT16x4_CMPGE_DEFINED +#endif + +#ifndef VUINT16x4_MIN_DEFINED +VEC_GENERIC_DBL_MIN(u, 16, 4, 2) +# define VUINT16x4_MIN_DEFINED +#endif + +#ifndef VUINT16x4_MAX_DEFINED +VEC_GENERIC_DBL_MAX(u, 16, 4, 2) +# define VUINT16x4_MAX_DEFINED +#endif + +#ifndef VUINT16x4_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(u, 16, 4, 2) +# define VUINT16x4_RSHIFT_DEFINED +#endif + +#ifndef VUINT16x4_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(u, 16, 4, 2) +# define VUINT16x4_LRSHIFT_DEFINED +#endif + +#ifndef VUINT16x4_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(u, 16, 4, 2) +# define VUINT16x4_LSHIFT_DEFINED +#endif + + + +/* vuint16x8 */ + +#ifndef VINT16x8_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(/* nothing */, 16, 8, 4) +# define VINT16x8_SPLAT_DEFINED +#endif + +#ifndef VINT16x8_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(/* nothing */, 16, 8, 4) +# define VINT16x8_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VINT16x8_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(/* nothing */, 16, 8, 4) +# define VINT16x8_LOAD_DEFINED +#endif + +#ifndef VINT16x8_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(/* nothing */, 16, 8, 4) +# define VINT16x8_STORE_ALIGNED_DEFINED +#endif + +#ifndef VINT16x8_STORE_DEFINED +VEC_GENERIC_DBL_STORE(/* nothing */, 16, 8, 4) +# define VINT16x8_STORE_DEFINED +#endif + +#ifndef VINT16x8_ADD_DEFINED +VEC_GENERIC_DBL_ADD(/* nothing */, 16, 8, 4) +# define VINT16x8_ADD_DEFINED +#endif + +#ifndef VINT16x8_SUB_DEFINED +VEC_GENERIC_DBL_SUB(/* nothing */, 16, 8, 4) +# define VINT16x8_SUB_DEFINED +#endif + +#ifndef VINT16x8_MUL_DEFINED +VEC_GENERIC_DBL_MUL(/* nothing */, 16, 8, 4) +# define VINT16x8_MUL_DEFINED +#endif + +#ifndef VINT16x8_DIV_DEFINED +VEC_GENERIC_DBL_DIV(/* nothing */, 16, 8, 4) +# define VINT16x8_DIV_DEFINED +#endif + +#ifndef VINT16x8_AVG_DEFINED +VEC_GENERIC_DBL_AVG(/* nothing */, 16, 8, 4) +# define VINT16x8_AVG_DEFINED +#endif + +#ifndef VINT16x8_AND_DEFINED +VEC_GENERIC_DBL_AND(/* nothing */, 16, 8, 4) +# define VINT16x8_AND_DEFINED +#endif + +#ifndef VINT16x8_OR_DEFINED +VEC_GENERIC_DBL_OR(/* nothing */, 16, 8, 4) +# define VINT16x8_OR_DEFINED +#endif + +#ifndef VINT16x8_XOR_DEFINED +VEC_GENERIC_DBL_XOR(/* nothing */, 16, 8, 4) +# define VINT16x8_XOR_DEFINED +#endif + +#ifndef VINT16x8_NOT_DEFINED +VEC_GENERIC_DBL_NOT(/* nothing */, 16, 8, 4) +# define VINT16x8_NOT_DEFINED +#endif + +#ifndef VINT16x8_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(/* nothing */, 16, 8, 4) +# define VINT16x8_CMPLT_DEFINED +#endif + +#ifndef VINT16x8_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(/* nothing */, 16, 8, 4) +# define VINT16x8_CMPEQ_DEFINED +#endif + +#ifndef VINT16x8_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(/* nothing */, 16, 8, 4) +# define VINT16x8_CMPGT_DEFINED +#endif + +#ifndef VINT16x8_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(/* nothing */, 16, 8, 4) +# define VINT16x8_CMPLE_DEFINED +#endif + +#ifndef VINT16x8_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(/* nothing */, 16, 8, 4) +# define VINT16x8_CMPGE_DEFINED +#endif + +#ifndef VINT16x8_MIN_DEFINED +VEC_GENERIC_DBL_MIN(/* nothing */, 16, 8, 4) +# define VINT16x8_MIN_DEFINED +#endif + +#ifndef VINT16x8_MAX_DEFINED +VEC_GENERIC_DBL_MAX(/* nothing */, 16, 8, 4) +# define VINT16x8_MAX_DEFINED +#endif + +#ifndef VINT16x8_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(/* nothing */, 16, 8, 4) +# define VINT16x8_RSHIFT_DEFINED +#endif + +#ifndef VINT16x8_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(/* nothing */, 16, 8, 4) +# define VINT16x8_LRSHIFT_DEFINED +#endif + +#ifndef VINT16x8_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(/* nothing */, 16, 8, 4) +# define VINT16x8_LSHIFT_DEFINED +#endif + + + +/* vint16x8 */ + +#ifndef VUINT16x8_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(u, 16, 8, 4) +# define VUINT16x8_SPLAT_DEFINED +#endif + +#ifndef VUINT16x8_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(u, 16, 8, 4) +# define VUINT16x8_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VUINT16x8_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(u, 16, 8, 4) +# define VUINT16x8_LOAD_DEFINED +#endif + +#ifndef VUINT16x8_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(u, 16, 8, 4) +# define VUINT16x8_STORE_ALIGNED_DEFINED +#endif + +#ifndef VUINT16x8_STORE_DEFINED +VEC_GENERIC_DBL_STORE(u, 16, 8, 4) +# define VUINT16x8_STORE_DEFINED +#endif + +#ifndef VUINT16x8_ADD_DEFINED +VEC_GENERIC_DBL_ADD(u, 16, 8, 4) +# define VUINT16x8_ADD_DEFINED +#endif + +#ifndef VUINT16x8_SUB_DEFINED +VEC_GENERIC_DBL_SUB(u, 16, 8, 4) +# define VUINT16x8_SUB_DEFINED +#endif + +#ifndef VUINT16x8_MUL_DEFINED +VEC_GENERIC_DBL_MUL(u, 16, 8, 4) +# define VUINT16x8_MUL_DEFINED +#endif + +#ifndef VUINT16x8_DIV_DEFINED +VEC_GENERIC_DBL_DIV(u, 16, 8, 4) +# define VUINT16x8_DIV_DEFINED +#endif + +#ifndef VUINT16x8_AVG_DEFINED +VEC_GENERIC_DBL_AVG(u, 16, 8, 4) +# define VUINT16x8_AVG_DEFINED +#endif + +#ifndef VUINT16x8_AND_DEFINED +VEC_GENERIC_DBL_AND(u, 16, 8, 4) +# define VUINT16x8_AND_DEFINED +#endif + +#ifndef VUINT16x8_OR_DEFINED +VEC_GENERIC_DBL_OR(u, 16, 8, 4) +# define VUINT16x8_OR_DEFINED +#endif + +#ifndef VUINT16x8_XOR_DEFINED +VEC_GENERIC_DBL_XOR(u, 16, 8, 4) +# define VUINT16x8_XOR_DEFINED +#endif + +#ifndef VUINT16x8_NOT_DEFINED +VEC_GENERIC_DBL_NOT(u, 16, 8, 4) +# define VUINT16x8_NOT_DEFINED +#endif + +#ifndef VUINT16x8_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(u, 16, 8, 4) +# define VUINT16x8_CMPLT_DEFINED +#endif + +#ifndef VUINT16x8_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(u, 16, 8, 4) +# define VUINT16x8_CMPEQ_DEFINED +#endif + +#ifndef VUINT16x8_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(u, 16, 8, 4) +# define VUINT16x8_CMPGT_DEFINED +#endif + +#ifndef VUINT16x8_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(u, 16, 8, 4) +# define VUINT16x8_CMPLE_DEFINED +#endif + +#ifndef VUINT16x8_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(u, 16, 8, 4) +# define VUINT16x8_CMPGE_DEFINED +#endif + +#ifndef VUINT16x8_MIN_DEFINED +VEC_GENERIC_DBL_MIN(u, 16, 8, 4) +# define VUINT16x8_MIN_DEFINED +#endif + +#ifndef VUINT16x8_MAX_DEFINED +VEC_GENERIC_DBL_MAX(u, 16, 8, 4) +# define VUINT16x8_MAX_DEFINED +#endif + +#ifndef VUINT16x8_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(u, 16, 8, 4) +# define VUINT16x8_RSHIFT_DEFINED +#endif + +#ifndef VUINT16x8_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(u, 16, 8, 4) +# define VUINT16x8_LRSHIFT_DEFINED +#endif + +#ifndef VUINT16x8_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(u, 16, 8, 4) +# define VUINT16x8_LSHIFT_DEFINED +#endif + + + +/* vuint16x16 */ + +#ifndef VINT16x16_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(/* nothing */, 16, 16, 8) +# define VINT16x16_SPLAT_DEFINED +#endif + +#ifndef VINT16x16_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(/* nothing */, 16, 16, 8) +# define VINT16x16_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VINT16x16_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(/* nothing */, 16, 16, 8) +# define VINT16x16_LOAD_DEFINED +#endif + +#ifndef VINT16x16_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(/* nothing */, 16, 16, 8) +# define VINT16x16_STORE_ALIGNED_DEFINED +#endif + +#ifndef VINT16x16_STORE_DEFINED +VEC_GENERIC_DBL_STORE(/* nothing */, 16, 16, 8) +# define VINT16x16_STORE_DEFINED +#endif + +#ifndef VINT16x16_ADD_DEFINED +VEC_GENERIC_DBL_ADD(/* nothing */, 16, 16, 8) +# define VINT16x16_ADD_DEFINED +#endif + +#ifndef VINT16x16_SUB_DEFINED +VEC_GENERIC_DBL_SUB(/* nothing */, 16, 16, 8) +# define VINT16x16_SUB_DEFINED +#endif + +#ifndef VINT16x16_MUL_DEFINED +VEC_GENERIC_DBL_MUL(/* nothing */, 16, 16, 8) +# define VINT16x16_MUL_DEFINED +#endif + +#ifndef VINT16x16_DIV_DEFINED +VEC_GENERIC_DBL_DIV(/* nothing */, 16, 16, 8) +# define VINT16x16_DIV_DEFINED +#endif + +#ifndef VINT16x16_AVG_DEFINED +VEC_GENERIC_DBL_AVG(/* nothing */, 16, 16, 8) +# define VINT16x16_AVG_DEFINED +#endif + +#ifndef VINT16x16_AND_DEFINED +VEC_GENERIC_DBL_AND(/* nothing */, 16, 16, 8) +# define VINT16x16_AND_DEFINED +#endif + +#ifndef VINT16x16_OR_DEFINED +VEC_GENERIC_DBL_OR(/* nothing */, 16, 16, 8) +# define VINT16x16_OR_DEFINED +#endif + +#ifndef VINT16x16_XOR_DEFINED +VEC_GENERIC_DBL_XOR(/* nothing */, 16, 16, 8) +# define VINT16x16_XOR_DEFINED +#endif + +#ifndef VINT16x16_NOT_DEFINED +VEC_GENERIC_DBL_NOT(/* nothing */, 16, 16, 8) +# define VINT16x16_NOT_DEFINED +#endif + +#ifndef VINT16x16_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(/* nothing */, 16, 16, 8) +# define VINT16x16_CMPLT_DEFINED +#endif + +#ifndef VINT16x16_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(/* nothing */, 16, 16, 8) +# define VINT16x16_CMPEQ_DEFINED +#endif + +#ifndef VINT16x16_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(/* nothing */, 16, 16, 8) +# define VINT16x16_CMPGT_DEFINED +#endif + +#ifndef VINT16x16_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(/* nothing */, 16, 16, 8) +# define VINT16x16_CMPLE_DEFINED +#endif + +#ifndef VINT16x16_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(/* nothing */, 16, 16, 8) +# define VINT16x16_CMPGE_DEFINED +#endif + +#ifndef VINT16x16_MIN_DEFINED +VEC_GENERIC_DBL_MIN(/* nothing */, 16, 16, 8) +# define VINT16x16_MIN_DEFINED +#endif + +#ifndef VINT16x16_MAX_DEFINED +VEC_GENERIC_DBL_MAX(/* nothing */, 16, 16, 8) +# define VINT16x16_MAX_DEFINED +#endif + +#ifndef VINT16x16_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(/* nothing */, 16, 16, 8) +# define VINT16x16_RSHIFT_DEFINED +#endif + +#ifndef VINT16x16_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(/* nothing */, 16, 16, 8) +# define VINT16x16_LRSHIFT_DEFINED +#endif + +#ifndef VINT16x16_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(/* nothing */, 16, 16, 8) +# define VINT16x16_LSHIFT_DEFINED +#endif + + + +/* vint16x16 */ + +#ifndef VUINT16x16_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(u, 16, 16, 8) +# define VUINT16x16_SPLAT_DEFINED +#endif + +#ifndef VUINT16x16_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(u, 16, 16, 8) +# define VUINT16x16_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VUINT16x16_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(u, 16, 16, 8) +# define VUINT16x16_LOAD_DEFINED +#endif + +#ifndef VUINT16x16_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(u, 16, 16, 8) +# define VUINT16x16_STORE_ALIGNED_DEFINED +#endif + +#ifndef VUINT16x16_STORE_DEFINED +VEC_GENERIC_DBL_STORE(u, 16, 16, 8) +# define VUINT16x16_STORE_DEFINED +#endif + +#ifndef VUINT16x16_ADD_DEFINED +VEC_GENERIC_DBL_ADD(u, 16, 16, 8) +# define VUINT16x16_ADD_DEFINED +#endif + +#ifndef VUINT16x16_SUB_DEFINED +VEC_GENERIC_DBL_SUB(u, 16, 16, 8) +# define VUINT16x16_SUB_DEFINED +#endif + +#ifndef VUINT16x16_MUL_DEFINED +VEC_GENERIC_DBL_MUL(u, 16, 16, 8) +# define VUINT16x16_MUL_DEFINED +#endif + +#ifndef VUINT16x16_DIV_DEFINED +VEC_GENERIC_DBL_DIV(u, 16, 16, 8) +# define VUINT16x16_DIV_DEFINED +#endif + +#ifndef VUINT16x16_AVG_DEFINED +VEC_GENERIC_DBL_AVG(u, 16, 16, 8) +# define VUINT16x16_AVG_DEFINED +#endif + +#ifndef VUINT16x16_AND_DEFINED +VEC_GENERIC_DBL_AND(u, 16, 16, 8) +# define VUINT16x16_AND_DEFINED +#endif + +#ifndef VUINT16x16_OR_DEFINED +VEC_GENERIC_DBL_OR(u, 16, 16, 8) +# define VUINT16x16_OR_DEFINED +#endif + +#ifndef VUINT16x16_XOR_DEFINED +VEC_GENERIC_DBL_XOR(u, 16, 16, 8) +# define VUINT16x16_XOR_DEFINED +#endif + +#ifndef VUINT16x16_NOT_DEFINED +VEC_GENERIC_DBL_NOT(u, 16, 16, 8) +# define VUINT16x16_NOT_DEFINED +#endif + +#ifndef VUINT16x16_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(u, 16, 16, 8) +# define VUINT16x16_CMPLT_DEFINED +#endif + +#ifndef VUINT16x16_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(u, 16, 16, 8) +# define VUINT16x16_CMPEQ_DEFINED +#endif + +#ifndef VUINT16x16_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(u, 16, 16, 8) +# define VUINT16x16_CMPGT_DEFINED +#endif + +#ifndef VUINT16x16_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(u, 16, 16, 8) +# define VUINT16x16_CMPLE_DEFINED +#endif + +#ifndef VUINT16x16_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(u, 16, 16, 8) +# define VUINT16x16_CMPGE_DEFINED +#endif + +#ifndef VUINT16x16_MIN_DEFINED +VEC_GENERIC_DBL_MIN(u, 16, 16, 8) +# define VUINT16x16_MIN_DEFINED +#endif + +#ifndef VUINT16x16_MAX_DEFINED +VEC_GENERIC_DBL_MAX(u, 16, 16, 8) +# define VUINT16x16_MAX_DEFINED +#endif + +#ifndef VUINT16x16_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(u, 16, 16, 8) +# define VUINT16x16_RSHIFT_DEFINED +#endif + +#ifndef VUINT16x16_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(u, 16, 16, 8) +# define VUINT16x16_LRSHIFT_DEFINED +#endif + +#ifndef VUINT16x16_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(u, 16, 16, 8) +# define VUINT16x16_LSHIFT_DEFINED +#endif + + + +/* vuint16x32 */ + +#ifndef VINT16x32_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(/* nothing */, 16, 32, 16) +# define VINT16x32_SPLAT_DEFINED +#endif + +#ifndef VINT16x32_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(/* nothing */, 16, 32, 16) +# define VINT16x32_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VINT16x32_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(/* nothing */, 16, 32, 16) +# define VINT16x32_LOAD_DEFINED +#endif + +#ifndef VINT16x32_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(/* nothing */, 16, 32, 16) +# define VINT16x32_STORE_ALIGNED_DEFINED +#endif + +#ifndef VINT16x32_STORE_DEFINED +VEC_GENERIC_DBL_STORE(/* nothing */, 16, 32, 16) +# define VINT16x32_STORE_DEFINED +#endif + +#ifndef VINT16x32_ADD_DEFINED +VEC_GENERIC_DBL_ADD(/* nothing */, 16, 32, 16) +# define VINT16x32_ADD_DEFINED +#endif + +#ifndef VINT16x32_SUB_DEFINED +VEC_GENERIC_DBL_SUB(/* nothing */, 16, 32, 16) +# define VINT16x32_SUB_DEFINED +#endif + +#ifndef VINT16x32_MUL_DEFINED +VEC_GENERIC_DBL_MUL(/* nothing */, 16, 32, 16) +# define VINT16x32_MUL_DEFINED +#endif + +#ifndef VINT16x32_DIV_DEFINED +VEC_GENERIC_DBL_DIV(/* nothing */, 16, 32, 16) +# define VINT16x32_DIV_DEFINED +#endif + +#ifndef VINT16x32_AVG_DEFINED +VEC_GENERIC_DBL_AVG(/* nothing */, 16, 32, 16) +# define VINT16x32_AVG_DEFINED +#endif + +#ifndef VINT16x32_AND_DEFINED +VEC_GENERIC_DBL_AND(/* nothing */, 16, 32, 16) +# define VINT16x32_AND_DEFINED +#endif + +#ifndef VINT16x32_OR_DEFINED +VEC_GENERIC_DBL_OR(/* nothing */, 16, 32, 16) +# define VINT16x32_OR_DEFINED +#endif + +#ifndef VINT16x32_XOR_DEFINED +VEC_GENERIC_DBL_XOR(/* nothing */, 16, 32, 16) +# define VINT16x32_XOR_DEFINED +#endif + +#ifndef VINT16x32_NOT_DEFINED +VEC_GENERIC_DBL_NOT(/* nothing */, 16, 32, 16) +# define VINT16x32_NOT_DEFINED +#endif + +#ifndef VINT16x32_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(/* nothing */, 16, 32, 16) +# define VINT16x32_CMPLT_DEFINED +#endif + +#ifndef VINT16x32_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(/* nothing */, 16, 32, 16) +# define VINT16x32_CMPEQ_DEFINED +#endif + +#ifndef VINT16x32_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(/* nothing */, 16, 32, 16) +# define VINT16x32_CMPGT_DEFINED +#endif + +#ifndef VINT16x32_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(/* nothing */, 16, 32, 16) +# define VINT16x32_CMPLE_DEFINED +#endif + +#ifndef VINT16x32_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(/* nothing */, 16, 32, 16) +# define VINT16x32_CMPGE_DEFINED +#endif + +#ifndef VINT16x32_MIN_DEFINED +VEC_GENERIC_DBL_MIN(/* nothing */, 16, 32, 16) +# define VINT16x32_MIN_DEFINED +#endif + +#ifndef VINT16x32_MAX_DEFINED +VEC_GENERIC_DBL_MAX(/* nothing */, 16, 32, 16) +# define VINT16x32_MAX_DEFINED +#endif + +#ifndef VINT16x32_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(/* nothing */, 16, 32, 16) +# define VINT16x32_RSHIFT_DEFINED +#endif + +#ifndef VINT16x32_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(/* nothing */, 16, 32, 16) +# define VINT16x32_LRSHIFT_DEFINED +#endif + +#ifndef VINT16x32_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(/* nothing */, 16, 32, 16) +# define VINT16x32_LSHIFT_DEFINED +#endif + + + +/* vint16x32 */ + +#ifndef VUINT16x32_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(u, 16, 32, 16) +# define VUINT16x32_SPLAT_DEFINED +#endif + +#ifndef VUINT16x32_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(u, 16, 32, 16) +# define VUINT16x32_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VUINT16x32_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(u, 16, 32, 16) +# define VUINT16x32_LOAD_DEFINED +#endif + +#ifndef VUINT16x32_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(u, 16, 32, 16) +# define VUINT16x32_STORE_ALIGNED_DEFINED +#endif + +#ifndef VUINT16x32_STORE_DEFINED +VEC_GENERIC_DBL_STORE(u, 16, 32, 16) +# define VUINT16x32_STORE_DEFINED +#endif + +#ifndef VUINT16x32_ADD_DEFINED +VEC_GENERIC_DBL_ADD(u, 16, 32, 16) +# define VUINT16x32_ADD_DEFINED +#endif + +#ifndef VUINT16x32_SUB_DEFINED +VEC_GENERIC_DBL_SUB(u, 16, 32, 16) +# define VUINT16x32_SUB_DEFINED +#endif + +#ifndef VUINT16x32_MUL_DEFINED +VEC_GENERIC_DBL_MUL(u, 16, 32, 16) +# define VUINT16x32_MUL_DEFINED +#endif + +#ifndef VUINT16x32_DIV_DEFINED +VEC_GENERIC_DBL_DIV(u, 16, 32, 16) +# define VUINT16x32_DIV_DEFINED +#endif + +#ifndef VUINT16x32_AVG_DEFINED +VEC_GENERIC_DBL_AVG(u, 16, 32, 16) +# define VUINT16x32_AVG_DEFINED +#endif + +#ifndef VUINT16x32_AND_DEFINED +VEC_GENERIC_DBL_AND(u, 16, 32, 16) +# define VUINT16x32_AND_DEFINED +#endif + +#ifndef VUINT16x32_OR_DEFINED +VEC_GENERIC_DBL_OR(u, 16, 32, 16) +# define VUINT16x32_OR_DEFINED +#endif + +#ifndef VUINT16x32_XOR_DEFINED +VEC_GENERIC_DBL_XOR(u, 16, 32, 16) +# define VUINT16x32_XOR_DEFINED +#endif + +#ifndef VUINT16x32_NOT_DEFINED +VEC_GENERIC_DBL_NOT(u, 16, 32, 16) +# define VUINT16x32_NOT_DEFINED +#endif + +#ifndef VUINT16x32_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(u, 16, 32, 16) +# define VUINT16x32_CMPLT_DEFINED +#endif + +#ifndef VUINT16x32_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(u, 16, 32, 16) +# define VUINT16x32_CMPEQ_DEFINED +#endif + +#ifndef VUINT16x32_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(u, 16, 32, 16) +# define VUINT16x32_CMPGT_DEFINED +#endif + +#ifndef VUINT16x32_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(u, 16, 32, 16) +# define VUINT16x32_CMPLE_DEFINED +#endif + +#ifndef VUINT16x32_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(u, 16, 32, 16) +# define VUINT16x32_CMPGE_DEFINED +#endif + +#ifndef VUINT16x32_MIN_DEFINED +VEC_GENERIC_DBL_MIN(u, 16, 32, 16) +# define VUINT16x32_MIN_DEFINED +#endif + +#ifndef VUINT16x32_MAX_DEFINED +VEC_GENERIC_DBL_MAX(u, 16, 32, 16) +# define VUINT16x32_MAX_DEFINED +#endif + +#ifndef VUINT16x32_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(u, 16, 32, 16) +# define VUINT16x32_RSHIFT_DEFINED +#endif + +#ifndef VUINT16x32_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(u, 16, 32, 16) +# define VUINT16x32_LRSHIFT_DEFINED +#endif + +#ifndef VUINT16x32_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(u, 16, 32, 16) +# define VUINT16x32_LSHIFT_DEFINED +#endif + + + +/* vuint32x2 */ + +#ifndef VINT32x2_SPLAT_DEFINED +VEC_GENERIC_SPLAT(/* nothing */, 32, 2) +# define VINT32x2_SPLAT_DEFINED +#endif +#ifndef VINT32x2_LOAD_ALIGNED_DEFINED +VEC_GENERIC_LOAD_ALIGNED(/* nothing */, 32, 2) +# define VINT32x2_LOAD_ALIGNED_DEFINED +#endif +#ifndef VINT32x2_LOAD_DEFINED +VEC_GENERIC_LOAD(/* nothing */, 32, 2) +# define VINT32x2_LOAD_DEFINED +#endif +#ifndef VINT32x2_STORE_ALIGNED_DEFINED +VEC_GENERIC_STORE_ALIGNED(/* nothing */, 32, 2) +# define VINT32x2_STORE_ALIGNED_DEFINED +#endif +#ifndef VINT32x2_STORE_DEFINED +VEC_GENERIC_STORE(/* nothing */, 32, 2) +# define VINT32x2_STORE_DEFINED +#endif +#ifndef VINT32x2_ADD_DEFINED +VEC_GENERIC_ADD(/* nothing */, 32, 2) +# define VINT32x2_ADD_DEFINED +#endif +#ifndef VINT32x2_SUB_DEFINED +VEC_GENERIC_SUB(/* nothing */, 32, 2) +# define VINT32x2_SUB_DEFINED +#endif +#ifndef VINT32x2_MUL_DEFINED +VEC_GENERIC_MUL(/* nothing */, 32, 2) +# define VINT32x2_MUL_DEFINED +#endif +#ifndef VINT32x2_DIV_DEFINED +VEC_GENERIC_DIV(/* nothing */, 32, 2) +# define VINT32x2_DIV_DEFINED +#endif +#ifndef VINT32x2_AVG_DEFINED +VEC_GENERIC_AVG(/* nothing */, 32, 2) +# define VINT32x2_AVG_DEFINED +#endif +#ifndef VINT32x2_AND_DEFINED +VEC_GENERIC_AND(/* nothing */, 32, 2) +# define VINT32x2_AND_DEFINED +#endif +#ifndef VINT32x2_OR_DEFINED +VEC_GENERIC_OR(/* nothing */, 32, 2) +# define VINT32x2_OR_DEFINED +#endif +#ifndef VINT32x2_XOR_DEFINED +VEC_GENERIC_XOR(/* nothing */, 32, 2) +# define VINT32x2_XOR_DEFINED +#endif +#ifndef VINT32x2_NOT_DEFINED +VEC_GENERIC_NOT(/* nothing */, 32, 2) +# define VINT32x2_NOT_DEFINED +#endif +#ifndef VINT32x2_CMPLT_DEFINED +VEC_GENERIC_CMPLT(/* nothing */, 32, 2) +# define VINT32x2_CMPLT_DEFINED +#endif +#ifndef VINT32x2_CMPEQ_DEFINED +VEC_GENERIC_CMPEQ(/* nothing */, 32, 2) +# define VINT32x2_CMPEQ_DEFINED +#endif +#ifndef VINT32x2_CMPGT_DEFINED +VEC_GENERIC_CMPGT(/* nothing */, 32, 2) +# define VINT32x2_CMPGT_DEFINED +#endif +#ifndef VINT32x2_CMPLE_DEFINED +VEC_GENERIC_CMPLE(/* nothing */, 32, 2) +# define VINT32x2_CMPLE_DEFINED +#endif +#ifndef VINT32x2_CMPGE_DEFINED +VEC_GENERIC_CMPGE(/* nothing */, 32, 2) +# define VINT32x2_CMPGE_DEFINED +#endif +#ifndef VINT32x2_MIN_DEFINED +VEC_GENERIC_MIN(/* nothing */, 32, 2) +# define VINT32x2_MIN_DEFINED +#endif +#ifndef VINT32x2_MAX_DEFINED +VEC_GENERIC_MAX(/* nothing */, 32, 2) +# define VINT32x2_MAX_DEFINED +#endif +#ifndef VINT32x2_RSHIFT_DEFINED +VEC_GENERIC_RSHIFT(/* nothing */, 32, 2) +# define VINT32x2_RSHIFT_DEFINED +#endif +#ifndef VINT32x2_LRSHIFT_DEFINED +VEC_GENERIC_LRSHIFT(/* nothing */, 32, 2) +# define VINT32x2_LRSHIFT_DEFINED +#endif +#ifndef VINT32x2_LSHIFT_DEFINED +VEC_GENERIC_LSHIFT(/* nothing */, 32, 2) +# define VINT32x2_LSHIFT_DEFINED +#endif + + +/* vint32x2 */ + +#ifndef VUINT32x2_SPLAT_DEFINED +VEC_GENERIC_SPLAT(u, 32, 2) +# define VUINT32x2_SPLAT_DEFINED +#endif +#ifndef VUINT32x2_LOAD_ALIGNED_DEFINED +VEC_GENERIC_LOAD_ALIGNED(u, 32, 2) +# define VUINT32x2_LOAD_ALIGNED_DEFINED +#endif +#ifndef VUINT32x2_LOAD_DEFINED +VEC_GENERIC_LOAD(u, 32, 2) +# define VUINT32x2_LOAD_DEFINED +#endif +#ifndef VUINT32x2_STORE_ALIGNED_DEFINED +VEC_GENERIC_STORE_ALIGNED(u, 32, 2) +# define VUINT32x2_STORE_ALIGNED_DEFINED +#endif +#ifndef VUINT32x2_STORE_DEFINED +VEC_GENERIC_STORE(u, 32, 2) +# define VUINT32x2_STORE_DEFINED +#endif +#ifndef VUINT32x2_ADD_DEFINED +VEC_GENERIC_ADD(u, 32, 2) +# define VUINT32x2_ADD_DEFINED +#endif +#ifndef VUINT32x2_SUB_DEFINED +VEC_GENERIC_SUB(u, 32, 2) +# define VUINT32x2_SUB_DEFINED +#endif +#ifndef VUINT32x2_MUL_DEFINED +VEC_GENERIC_MUL(u, 32, 2) +# define VUINT32x2_MUL_DEFINED +#endif +#ifndef VUINT32x2_DIV_DEFINED +VEC_GENERIC_DIV(u, 32, 2) +# define VUINT32x2_DIV_DEFINED +#endif +#ifndef VUINT32x2_AVG_DEFINED +VEC_GENERIC_AVG(u, 32, 2) +# define VUINT32x2_AVG_DEFINED +#endif +#ifndef VUINT32x2_AND_DEFINED +VEC_GENERIC_AND(u, 32, 2) +# define VUINT32x2_AND_DEFINED +#endif +#ifndef VUINT32x2_OR_DEFINED +VEC_GENERIC_OR(u, 32, 2) +# define VUINT32x2_OR_DEFINED +#endif +#ifndef VUINT32x2_XOR_DEFINED +VEC_GENERIC_XOR(u, 32, 2) +# define VUINT32x2_XOR_DEFINED +#endif +#ifndef VUINT32x2_NOT_DEFINED +VEC_GENERIC_NOT(u, 32, 2) +# define VUINT32x2_NOT_DEFINED +#endif +#ifndef VUINT32x2_CMPLT_DEFINED +VEC_GENERIC_CMPLT(u, 32, 2) +# define VUINT32x2_CMPLT_DEFINED +#endif +#ifndef VUINT32x2_CMPEQ_DEFINED +VEC_GENERIC_CMPEQ(u, 32, 2) +# define VUINT32x2_CMPEQ_DEFINED +#endif +#ifndef VUINT32x2_CMPGT_DEFINED +VEC_GENERIC_CMPGT(u, 32, 2) +# define VUINT32x2_CMPGT_DEFINED +#endif +#ifndef VUINT32x2_CMPLE_DEFINED +VEC_GENERIC_CMPLE(u, 32, 2) +# define VUINT32x2_CMPLE_DEFINED +#endif +#ifndef VUINT32x2_CMPGE_DEFINED +VEC_GENERIC_CMPGE(u, 32, 2) +# define VUINT32x2_CMPGE_DEFINED +#endif +#ifndef VUINT32x2_MIN_DEFINED +VEC_GENERIC_MIN(u, 32, 2) +# define VUINT32x2_MIN_DEFINED +#endif +#ifndef VUINT32x2_MAX_DEFINED +VEC_GENERIC_MAX(u, 32, 2) +# define VUINT32x2_MAX_DEFINED +#endif +#ifndef VUINT32x2_RSHIFT_DEFINED +VEC_GENERIC_RSHIFT(u, 32, 2) +# define VUINT32x2_RSHIFT_DEFINED +#endif +#ifndef VUINT32x2_LRSHIFT_DEFINED +VEC_GENERIC_LRSHIFT(u, 32, 2) +# define VUINT32x2_LRSHIFT_DEFINED +#endif +#ifndef VUINT32x2_LSHIFT_DEFINED +VEC_GENERIC_LSHIFT(u, 32, 2) +# define VUINT32x2_LSHIFT_DEFINED +#endif + + +/* vuint32x4 */ + +#ifndef VINT32x4_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(/* nothing */, 32, 4, 2) +# define VINT32x4_SPLAT_DEFINED +#endif + +#ifndef VINT32x4_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(/* nothing */, 32, 4, 2) +# define VINT32x4_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VINT32x4_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(/* nothing */, 32, 4, 2) +# define VINT32x4_LOAD_DEFINED +#endif + +#ifndef VINT32x4_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(/* nothing */, 32, 4, 2) +# define VINT32x4_STORE_ALIGNED_DEFINED +#endif + +#ifndef VINT32x4_STORE_DEFINED +VEC_GENERIC_DBL_STORE(/* nothing */, 32, 4, 2) +# define VINT32x4_STORE_DEFINED +#endif + +#ifndef VINT32x4_ADD_DEFINED +VEC_GENERIC_DBL_ADD(/* nothing */, 32, 4, 2) +# define VINT32x4_ADD_DEFINED +#endif + +#ifndef VINT32x4_SUB_DEFINED +VEC_GENERIC_DBL_SUB(/* nothing */, 32, 4, 2) +# define VINT32x4_SUB_DEFINED +#endif + +#ifndef VINT32x4_MUL_DEFINED +VEC_GENERIC_DBL_MUL(/* nothing */, 32, 4, 2) +# define VINT32x4_MUL_DEFINED +#endif + +#ifndef VINT32x4_DIV_DEFINED +VEC_GENERIC_DBL_DIV(/* nothing */, 32, 4, 2) +# define VINT32x4_DIV_DEFINED +#endif + +#ifndef VINT32x4_AVG_DEFINED +VEC_GENERIC_DBL_AVG(/* nothing */, 32, 4, 2) +# define VINT32x4_AVG_DEFINED +#endif + +#ifndef VINT32x4_AND_DEFINED +VEC_GENERIC_DBL_AND(/* nothing */, 32, 4, 2) +# define VINT32x4_AND_DEFINED +#endif + +#ifndef VINT32x4_OR_DEFINED +VEC_GENERIC_DBL_OR(/* nothing */, 32, 4, 2) +# define VINT32x4_OR_DEFINED +#endif + +#ifndef VINT32x4_XOR_DEFINED +VEC_GENERIC_DBL_XOR(/* nothing */, 32, 4, 2) +# define VINT32x4_XOR_DEFINED +#endif + +#ifndef VINT32x4_NOT_DEFINED +VEC_GENERIC_DBL_NOT(/* nothing */, 32, 4, 2) +# define VINT32x4_NOT_DEFINED +#endif + +#ifndef VINT32x4_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(/* nothing */, 32, 4, 2) +# define VINT32x4_CMPLT_DEFINED +#endif + +#ifndef VINT32x4_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(/* nothing */, 32, 4, 2) +# define VINT32x4_CMPEQ_DEFINED +#endif + +#ifndef VINT32x4_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(/* nothing */, 32, 4, 2) +# define VINT32x4_CMPGT_DEFINED +#endif + +#ifndef VINT32x4_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(/* nothing */, 32, 4, 2) +# define VINT32x4_CMPLE_DEFINED +#endif + +#ifndef VINT32x4_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(/* nothing */, 32, 4, 2) +# define VINT32x4_CMPGE_DEFINED +#endif + +#ifndef VINT32x4_MIN_DEFINED +VEC_GENERIC_DBL_MIN(/* nothing */, 32, 4, 2) +# define VINT32x4_MIN_DEFINED +#endif + +#ifndef VINT32x4_MAX_DEFINED +VEC_GENERIC_DBL_MAX(/* nothing */, 32, 4, 2) +# define VINT32x4_MAX_DEFINED +#endif + +#ifndef VINT32x4_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(/* nothing */, 32, 4, 2) +# define VINT32x4_RSHIFT_DEFINED +#endif + +#ifndef VINT32x4_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(/* nothing */, 32, 4, 2) +# define VINT32x4_LRSHIFT_DEFINED +#endif + +#ifndef VINT32x4_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(/* nothing */, 32, 4, 2) +# define VINT32x4_LSHIFT_DEFINED +#endif + + + +/* vint32x4 */ + +#ifndef VUINT32x4_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(u, 32, 4, 2) +# define VUINT32x4_SPLAT_DEFINED +#endif + +#ifndef VUINT32x4_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(u, 32, 4, 2) +# define VUINT32x4_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VUINT32x4_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(u, 32, 4, 2) +# define VUINT32x4_LOAD_DEFINED +#endif + +#ifndef VUINT32x4_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(u, 32, 4, 2) +# define VUINT32x4_STORE_ALIGNED_DEFINED +#endif + +#ifndef VUINT32x4_STORE_DEFINED +VEC_GENERIC_DBL_STORE(u, 32, 4, 2) +# define VUINT32x4_STORE_DEFINED +#endif + +#ifndef VUINT32x4_ADD_DEFINED +VEC_GENERIC_DBL_ADD(u, 32, 4, 2) +# define VUINT32x4_ADD_DEFINED +#endif + +#ifndef VUINT32x4_SUB_DEFINED +VEC_GENERIC_DBL_SUB(u, 32, 4, 2) +# define VUINT32x4_SUB_DEFINED +#endif + +#ifndef VUINT32x4_MUL_DEFINED +VEC_GENERIC_DBL_MUL(u, 32, 4, 2) +# define VUINT32x4_MUL_DEFINED +#endif + +#ifndef VUINT32x4_DIV_DEFINED +VEC_GENERIC_DBL_DIV(u, 32, 4, 2) +# define VUINT32x4_DIV_DEFINED +#endif + +#ifndef VUINT32x4_AVG_DEFINED +VEC_GENERIC_DBL_AVG(u, 32, 4, 2) +# define VUINT32x4_AVG_DEFINED +#endif + +#ifndef VUINT32x4_AND_DEFINED +VEC_GENERIC_DBL_AND(u, 32, 4, 2) +# define VUINT32x4_AND_DEFINED +#endif + +#ifndef VUINT32x4_OR_DEFINED +VEC_GENERIC_DBL_OR(u, 32, 4, 2) +# define VUINT32x4_OR_DEFINED +#endif + +#ifndef VUINT32x4_XOR_DEFINED +VEC_GENERIC_DBL_XOR(u, 32, 4, 2) +# define VUINT32x4_XOR_DEFINED +#endif + +#ifndef VUINT32x4_NOT_DEFINED +VEC_GENERIC_DBL_NOT(u, 32, 4, 2) +# define VUINT32x4_NOT_DEFINED +#endif + +#ifndef VUINT32x4_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(u, 32, 4, 2) +# define VUINT32x4_CMPLT_DEFINED +#endif + +#ifndef VUINT32x4_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(u, 32, 4, 2) +# define VUINT32x4_CMPEQ_DEFINED +#endif + +#ifndef VUINT32x4_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(u, 32, 4, 2) +# define VUINT32x4_CMPGT_DEFINED +#endif + +#ifndef VUINT32x4_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(u, 32, 4, 2) +# define VUINT32x4_CMPLE_DEFINED +#endif + +#ifndef VUINT32x4_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(u, 32, 4, 2) +# define VUINT32x4_CMPGE_DEFINED +#endif + +#ifndef VUINT32x4_MIN_DEFINED +VEC_GENERIC_DBL_MIN(u, 32, 4, 2) +# define VUINT32x4_MIN_DEFINED +#endif + +#ifndef VUINT32x4_MAX_DEFINED +VEC_GENERIC_DBL_MAX(u, 32, 4, 2) +# define VUINT32x4_MAX_DEFINED +#endif + +#ifndef VUINT32x4_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(u, 32, 4, 2) +# define VUINT32x4_RSHIFT_DEFINED +#endif + +#ifndef VUINT32x4_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(u, 32, 4, 2) +# define VUINT32x4_LRSHIFT_DEFINED +#endif + +#ifndef VUINT32x4_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(u, 32, 4, 2) +# define VUINT32x4_LSHIFT_DEFINED +#endif + + + +/* vuint32x8 */ + +#ifndef VINT32x8_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(/* nothing */, 32, 8, 4) +# define VINT32x8_SPLAT_DEFINED +#endif + +#ifndef VINT32x8_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(/* nothing */, 32, 8, 4) +# define VINT32x8_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VINT32x8_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(/* nothing */, 32, 8, 4) +# define VINT32x8_LOAD_DEFINED +#endif + +#ifndef VINT32x8_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(/* nothing */, 32, 8, 4) +# define VINT32x8_STORE_ALIGNED_DEFINED +#endif + +#ifndef VINT32x8_STORE_DEFINED +VEC_GENERIC_DBL_STORE(/* nothing */, 32, 8, 4) +# define VINT32x8_STORE_DEFINED +#endif + +#ifndef VINT32x8_ADD_DEFINED +VEC_GENERIC_DBL_ADD(/* nothing */, 32, 8, 4) +# define VINT32x8_ADD_DEFINED +#endif + +#ifndef VINT32x8_SUB_DEFINED +VEC_GENERIC_DBL_SUB(/* nothing */, 32, 8, 4) +# define VINT32x8_SUB_DEFINED +#endif + +#ifndef VINT32x8_MUL_DEFINED +VEC_GENERIC_DBL_MUL(/* nothing */, 32, 8, 4) +# define VINT32x8_MUL_DEFINED +#endif + +#ifndef VINT32x8_DIV_DEFINED +VEC_GENERIC_DBL_DIV(/* nothing */, 32, 8, 4) +# define VINT32x8_DIV_DEFINED +#endif + +#ifndef VINT32x8_AVG_DEFINED +VEC_GENERIC_DBL_AVG(/* nothing */, 32, 8, 4) +# define VINT32x8_AVG_DEFINED +#endif + +#ifndef VINT32x8_AND_DEFINED +VEC_GENERIC_DBL_AND(/* nothing */, 32, 8, 4) +# define VINT32x8_AND_DEFINED +#endif + +#ifndef VINT32x8_OR_DEFINED +VEC_GENERIC_DBL_OR(/* nothing */, 32, 8, 4) +# define VINT32x8_OR_DEFINED +#endif + +#ifndef VINT32x8_XOR_DEFINED +VEC_GENERIC_DBL_XOR(/* nothing */, 32, 8, 4) +# define VINT32x8_XOR_DEFINED +#endif + +#ifndef VINT32x8_NOT_DEFINED +VEC_GENERIC_DBL_NOT(/* nothing */, 32, 8, 4) +# define VINT32x8_NOT_DEFINED +#endif + +#ifndef VINT32x8_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(/* nothing */, 32, 8, 4) +# define VINT32x8_CMPLT_DEFINED +#endif + +#ifndef VINT32x8_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(/* nothing */, 32, 8, 4) +# define VINT32x8_CMPEQ_DEFINED +#endif + +#ifndef VINT32x8_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(/* nothing */, 32, 8, 4) +# define VINT32x8_CMPGT_DEFINED +#endif + +#ifndef VINT32x8_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(/* nothing */, 32, 8, 4) +# define VINT32x8_CMPLE_DEFINED +#endif + +#ifndef VINT32x8_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(/* nothing */, 32, 8, 4) +# define VINT32x8_CMPGE_DEFINED +#endif + +#ifndef VINT32x8_MIN_DEFINED +VEC_GENERIC_DBL_MIN(/* nothing */, 32, 8, 4) +# define VINT32x8_MIN_DEFINED +#endif + +#ifndef VINT32x8_MAX_DEFINED +VEC_GENERIC_DBL_MAX(/* nothing */, 32, 8, 4) +# define VINT32x8_MAX_DEFINED +#endif + +#ifndef VINT32x8_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(/* nothing */, 32, 8, 4) +# define VINT32x8_RSHIFT_DEFINED +#endif + +#ifndef VINT32x8_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(/* nothing */, 32, 8, 4) +# define VINT32x8_LRSHIFT_DEFINED +#endif + +#ifndef VINT32x8_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(/* nothing */, 32, 8, 4) +# define VINT32x8_LSHIFT_DEFINED +#endif + + + +/* vint32x8 */ + +#ifndef VUINT32x8_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(u, 32, 8, 4) +# define VUINT32x8_SPLAT_DEFINED +#endif + +#ifndef VUINT32x8_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(u, 32, 8, 4) +# define VUINT32x8_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VUINT32x8_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(u, 32, 8, 4) +# define VUINT32x8_LOAD_DEFINED +#endif + +#ifndef VUINT32x8_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(u, 32, 8, 4) +# define VUINT32x8_STORE_ALIGNED_DEFINED +#endif + +#ifndef VUINT32x8_STORE_DEFINED +VEC_GENERIC_DBL_STORE(u, 32, 8, 4) +# define VUINT32x8_STORE_DEFINED +#endif + +#ifndef VUINT32x8_ADD_DEFINED +VEC_GENERIC_DBL_ADD(u, 32, 8, 4) +# define VUINT32x8_ADD_DEFINED +#endif + +#ifndef VUINT32x8_SUB_DEFINED +VEC_GENERIC_DBL_SUB(u, 32, 8, 4) +# define VUINT32x8_SUB_DEFINED +#endif + +#ifndef VUINT32x8_MUL_DEFINED +VEC_GENERIC_DBL_MUL(u, 32, 8, 4) +# define VUINT32x8_MUL_DEFINED +#endif + +#ifndef VUINT32x8_DIV_DEFINED +VEC_GENERIC_DBL_DIV(u, 32, 8, 4) +# define VUINT32x8_DIV_DEFINED +#endif + +#ifndef VUINT32x8_AVG_DEFINED +VEC_GENERIC_DBL_AVG(u, 32, 8, 4) +# define VUINT32x8_AVG_DEFINED +#endif + +#ifndef VUINT32x8_AND_DEFINED +VEC_GENERIC_DBL_AND(u, 32, 8, 4) +# define VUINT32x8_AND_DEFINED +#endif + +#ifndef VUINT32x8_OR_DEFINED +VEC_GENERIC_DBL_OR(u, 32, 8, 4) +# define VUINT32x8_OR_DEFINED +#endif + +#ifndef VUINT32x8_XOR_DEFINED +VEC_GENERIC_DBL_XOR(u, 32, 8, 4) +# define VUINT32x8_XOR_DEFINED +#endif + +#ifndef VUINT32x8_NOT_DEFINED +VEC_GENERIC_DBL_NOT(u, 32, 8, 4) +# define VUINT32x8_NOT_DEFINED +#endif + +#ifndef VUINT32x8_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(u, 32, 8, 4) +# define VUINT32x8_CMPLT_DEFINED +#endif + +#ifndef VUINT32x8_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(u, 32, 8, 4) +# define VUINT32x8_CMPEQ_DEFINED +#endif + +#ifndef VUINT32x8_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(u, 32, 8, 4) +# define VUINT32x8_CMPGT_DEFINED +#endif + +#ifndef VUINT32x8_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(u, 32, 8, 4) +# define VUINT32x8_CMPLE_DEFINED +#endif + +#ifndef VUINT32x8_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(u, 32, 8, 4) +# define VUINT32x8_CMPGE_DEFINED +#endif + +#ifndef VUINT32x8_MIN_DEFINED +VEC_GENERIC_DBL_MIN(u, 32, 8, 4) +# define VUINT32x8_MIN_DEFINED +#endif + +#ifndef VUINT32x8_MAX_DEFINED +VEC_GENERIC_DBL_MAX(u, 32, 8, 4) +# define VUINT32x8_MAX_DEFINED +#endif + +#ifndef VUINT32x8_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(u, 32, 8, 4) +# define VUINT32x8_RSHIFT_DEFINED +#endif + +#ifndef VUINT32x8_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(u, 32, 8, 4) +# define VUINT32x8_LRSHIFT_DEFINED +#endif + +#ifndef VUINT32x8_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(u, 32, 8, 4) +# define VUINT32x8_LSHIFT_DEFINED +#endif + + + +/* vuint32x16 */ + +#ifndef VINT32x16_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(/* nothing */, 32, 16, 8) +# define VINT32x16_SPLAT_DEFINED +#endif + +#ifndef VINT32x16_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(/* nothing */, 32, 16, 8) +# define VINT32x16_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VINT32x16_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(/* nothing */, 32, 16, 8) +# define VINT32x16_LOAD_DEFINED +#endif + +#ifndef VINT32x16_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(/* nothing */, 32, 16, 8) +# define VINT32x16_STORE_ALIGNED_DEFINED +#endif + +#ifndef VINT32x16_STORE_DEFINED +VEC_GENERIC_DBL_STORE(/* nothing */, 32, 16, 8) +# define VINT32x16_STORE_DEFINED +#endif + +#ifndef VINT32x16_ADD_DEFINED +VEC_GENERIC_DBL_ADD(/* nothing */, 32, 16, 8) +# define VINT32x16_ADD_DEFINED +#endif + +#ifndef VINT32x16_SUB_DEFINED +VEC_GENERIC_DBL_SUB(/* nothing */, 32, 16, 8) +# define VINT32x16_SUB_DEFINED +#endif + +#ifndef VINT32x16_MUL_DEFINED +VEC_GENERIC_DBL_MUL(/* nothing */, 32, 16, 8) +# define VINT32x16_MUL_DEFINED +#endif + +#ifndef VINT32x16_DIV_DEFINED +VEC_GENERIC_DBL_DIV(/* nothing */, 32, 16, 8) +# define VINT32x16_DIV_DEFINED +#endif + +#ifndef VINT32x16_AVG_DEFINED +VEC_GENERIC_DBL_AVG(/* nothing */, 32, 16, 8) +# define VINT32x16_AVG_DEFINED +#endif + +#ifndef VINT32x16_AND_DEFINED +VEC_GENERIC_DBL_AND(/* nothing */, 32, 16, 8) +# define VINT32x16_AND_DEFINED +#endif + +#ifndef VINT32x16_OR_DEFINED +VEC_GENERIC_DBL_OR(/* nothing */, 32, 16, 8) +# define VINT32x16_OR_DEFINED +#endif + +#ifndef VINT32x16_XOR_DEFINED +VEC_GENERIC_DBL_XOR(/* nothing */, 32, 16, 8) +# define VINT32x16_XOR_DEFINED +#endif + +#ifndef VINT32x16_NOT_DEFINED +VEC_GENERIC_DBL_NOT(/* nothing */, 32, 16, 8) +# define VINT32x16_NOT_DEFINED +#endif + +#ifndef VINT32x16_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(/* nothing */, 32, 16, 8) +# define VINT32x16_CMPLT_DEFINED +#endif + +#ifndef VINT32x16_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(/* nothing */, 32, 16, 8) +# define VINT32x16_CMPEQ_DEFINED +#endif + +#ifndef VINT32x16_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(/* nothing */, 32, 16, 8) +# define VINT32x16_CMPGT_DEFINED +#endif + +#ifndef VINT32x16_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(/* nothing */, 32, 16, 8) +# define VINT32x16_CMPLE_DEFINED +#endif + +#ifndef VINT32x16_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(/* nothing */, 32, 16, 8) +# define VINT32x16_CMPGE_DEFINED +#endif + +#ifndef VINT32x16_MIN_DEFINED +VEC_GENERIC_DBL_MIN(/* nothing */, 32, 16, 8) +# define VINT32x16_MIN_DEFINED +#endif + +#ifndef VINT32x16_MAX_DEFINED +VEC_GENERIC_DBL_MAX(/* nothing */, 32, 16, 8) +# define VINT32x16_MAX_DEFINED +#endif + +#ifndef VINT32x16_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(/* nothing */, 32, 16, 8) +# define VINT32x16_RSHIFT_DEFINED +#endif + +#ifndef VINT32x16_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(/* nothing */, 32, 16, 8) +# define VINT32x16_LRSHIFT_DEFINED +#endif + +#ifndef VINT32x16_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(/* nothing */, 32, 16, 8) +# define VINT32x16_LSHIFT_DEFINED +#endif + + + +/* vint32x16 */ + +#ifndef VUINT32x16_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(u, 32, 16, 8) +# define VUINT32x16_SPLAT_DEFINED +#endif + +#ifndef VUINT32x16_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(u, 32, 16, 8) +# define VUINT32x16_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VUINT32x16_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(u, 32, 16, 8) +# define VUINT32x16_LOAD_DEFINED +#endif + +#ifndef VUINT32x16_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(u, 32, 16, 8) +# define VUINT32x16_STORE_ALIGNED_DEFINED +#endif + +#ifndef VUINT32x16_STORE_DEFINED +VEC_GENERIC_DBL_STORE(u, 32, 16, 8) +# define VUINT32x16_STORE_DEFINED +#endif + +#ifndef VUINT32x16_ADD_DEFINED +VEC_GENERIC_DBL_ADD(u, 32, 16, 8) +# define VUINT32x16_ADD_DEFINED +#endif + +#ifndef VUINT32x16_SUB_DEFINED +VEC_GENERIC_DBL_SUB(u, 32, 16, 8) +# define VUINT32x16_SUB_DEFINED +#endif + +#ifndef VUINT32x16_MUL_DEFINED +VEC_GENERIC_DBL_MUL(u, 32, 16, 8) +# define VUINT32x16_MUL_DEFINED +#endif + +#ifndef VUINT32x16_DIV_DEFINED +VEC_GENERIC_DBL_DIV(u, 32, 16, 8) +# define VUINT32x16_DIV_DEFINED +#endif + +#ifndef VUINT32x16_AVG_DEFINED +VEC_GENERIC_DBL_AVG(u, 32, 16, 8) +# define VUINT32x16_AVG_DEFINED +#endif + +#ifndef VUINT32x16_AND_DEFINED +VEC_GENERIC_DBL_AND(u, 32, 16, 8) +# define VUINT32x16_AND_DEFINED +#endif + +#ifndef VUINT32x16_OR_DEFINED +VEC_GENERIC_DBL_OR(u, 32, 16, 8) +# define VUINT32x16_OR_DEFINED +#endif + +#ifndef VUINT32x16_XOR_DEFINED +VEC_GENERIC_DBL_XOR(u, 32, 16, 8) +# define VUINT32x16_XOR_DEFINED +#endif + +#ifndef VUINT32x16_NOT_DEFINED +VEC_GENERIC_DBL_NOT(u, 32, 16, 8) +# define VUINT32x16_NOT_DEFINED +#endif + +#ifndef VUINT32x16_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(u, 32, 16, 8) +# define VUINT32x16_CMPLT_DEFINED +#endif + +#ifndef VUINT32x16_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(u, 32, 16, 8) +# define VUINT32x16_CMPEQ_DEFINED +#endif + +#ifndef VUINT32x16_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(u, 32, 16, 8) +# define VUINT32x16_CMPGT_DEFINED +#endif + +#ifndef VUINT32x16_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(u, 32, 16, 8) +# define VUINT32x16_CMPLE_DEFINED +#endif + +#ifndef VUINT32x16_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(u, 32, 16, 8) +# define VUINT32x16_CMPGE_DEFINED +#endif + +#ifndef VUINT32x16_MIN_DEFINED +VEC_GENERIC_DBL_MIN(u, 32, 16, 8) +# define VUINT32x16_MIN_DEFINED +#endif + +#ifndef VUINT32x16_MAX_DEFINED +VEC_GENERIC_DBL_MAX(u, 32, 16, 8) +# define VUINT32x16_MAX_DEFINED +#endif + +#ifndef VUINT32x16_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(u, 32, 16, 8) +# define VUINT32x16_RSHIFT_DEFINED +#endif + +#ifndef VUINT32x16_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(u, 32, 16, 8) +# define VUINT32x16_LRSHIFT_DEFINED +#endif + +#ifndef VUINT32x16_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(u, 32, 16, 8) +# define VUINT32x16_LSHIFT_DEFINED +#endif + + + +/* vuint64x2 */ + +#ifndef VINT64x2_SPLAT_DEFINED +VEC_GENERIC_SPLAT(/* nothing */, 64, 2) +# define VINT64x2_SPLAT_DEFINED +#endif +#ifndef VINT64x2_LOAD_ALIGNED_DEFINED +VEC_GENERIC_LOAD_ALIGNED(/* nothing */, 64, 2) +# define VINT64x2_LOAD_ALIGNED_DEFINED +#endif +#ifndef VINT64x2_LOAD_DEFINED +VEC_GENERIC_LOAD(/* nothing */, 64, 2) +# define VINT64x2_LOAD_DEFINED +#endif +#ifndef VINT64x2_STORE_ALIGNED_DEFINED +VEC_GENERIC_STORE_ALIGNED(/* nothing */, 64, 2) +# define VINT64x2_STORE_ALIGNED_DEFINED +#endif +#ifndef VINT64x2_STORE_DEFINED +VEC_GENERIC_STORE(/* nothing */, 64, 2) +# define VINT64x2_STORE_DEFINED +#endif +#ifndef VINT64x2_ADD_DEFINED +VEC_GENERIC_ADD(/* nothing */, 64, 2) +# define VINT64x2_ADD_DEFINED +#endif +#ifndef VINT64x2_SUB_DEFINED +VEC_GENERIC_SUB(/* nothing */, 64, 2) +# define VINT64x2_SUB_DEFINED +#endif +#ifndef VINT64x2_MUL_DEFINED +VEC_GENERIC_MUL(/* nothing */, 64, 2) +# define VINT64x2_MUL_DEFINED +#endif +#ifndef VINT64x2_DIV_DEFINED +VEC_GENERIC_DIV(/* nothing */, 64, 2) +# define VINT64x2_DIV_DEFINED +#endif +#ifndef VINT64x2_AVG_DEFINED +VEC_GENERIC_AVG(/* nothing */, 64, 2) +# define VINT64x2_AVG_DEFINED +#endif +#ifndef VINT64x2_AND_DEFINED +VEC_GENERIC_AND(/* nothing */, 64, 2) +# define VINT64x2_AND_DEFINED +#endif +#ifndef VINT64x2_OR_DEFINED +VEC_GENERIC_OR(/* nothing */, 64, 2) +# define VINT64x2_OR_DEFINED +#endif +#ifndef VINT64x2_XOR_DEFINED +VEC_GENERIC_XOR(/* nothing */, 64, 2) +# define VINT64x2_XOR_DEFINED +#endif +#ifndef VINT64x2_NOT_DEFINED +VEC_GENERIC_NOT(/* nothing */, 64, 2) +# define VINT64x2_NOT_DEFINED +#endif +#ifndef VINT64x2_CMPLT_DEFINED +VEC_GENERIC_CMPLT(/* nothing */, 64, 2) +# define VINT64x2_CMPLT_DEFINED +#endif +#ifndef VINT64x2_CMPEQ_DEFINED +VEC_GENERIC_CMPEQ(/* nothing */, 64, 2) +# define VINT64x2_CMPEQ_DEFINED +#endif +#ifndef VINT64x2_CMPGT_DEFINED +VEC_GENERIC_CMPGT(/* nothing */, 64, 2) +# define VINT64x2_CMPGT_DEFINED +#endif +#ifndef VINT64x2_CMPLE_DEFINED +VEC_GENERIC_CMPLE(/* nothing */, 64, 2) +# define VINT64x2_CMPLE_DEFINED +#endif +#ifndef VINT64x2_CMPGE_DEFINED +VEC_GENERIC_CMPGE(/* nothing */, 64, 2) +# define VINT64x2_CMPGE_DEFINED +#endif +#ifndef VINT64x2_MIN_DEFINED +VEC_GENERIC_MIN(/* nothing */, 64, 2) +# define VINT64x2_MIN_DEFINED +#endif +#ifndef VINT64x2_MAX_DEFINED +VEC_GENERIC_MAX(/* nothing */, 64, 2) +# define VINT64x2_MAX_DEFINED +#endif +#ifndef VINT64x2_RSHIFT_DEFINED +VEC_GENERIC_RSHIFT(/* nothing */, 64, 2) +# define VINT64x2_RSHIFT_DEFINED +#endif +#ifndef VINT64x2_LRSHIFT_DEFINED +VEC_GENERIC_LRSHIFT(/* nothing */, 64, 2) +# define VINT64x2_LRSHIFT_DEFINED +#endif +#ifndef VINT64x2_LSHIFT_DEFINED +VEC_GENERIC_LSHIFT(/* nothing */, 64, 2) +# define VINT64x2_LSHIFT_DEFINED +#endif + + +/* vint64x2 */ + +#ifndef VUINT64x2_SPLAT_DEFINED +VEC_GENERIC_SPLAT(u, 64, 2) +# define VUINT64x2_SPLAT_DEFINED +#endif +#ifndef VUINT64x2_LOAD_ALIGNED_DEFINED +VEC_GENERIC_LOAD_ALIGNED(u, 64, 2) +# define VUINT64x2_LOAD_ALIGNED_DEFINED +#endif +#ifndef VUINT64x2_LOAD_DEFINED +VEC_GENERIC_LOAD(u, 64, 2) +# define VUINT64x2_LOAD_DEFINED +#endif +#ifndef VUINT64x2_STORE_ALIGNED_DEFINED +VEC_GENERIC_STORE_ALIGNED(u, 64, 2) +# define VUINT64x2_STORE_ALIGNED_DEFINED +#endif +#ifndef VUINT64x2_STORE_DEFINED +VEC_GENERIC_STORE(u, 64, 2) +# define VUINT64x2_STORE_DEFINED +#endif +#ifndef VUINT64x2_ADD_DEFINED +VEC_GENERIC_ADD(u, 64, 2) +# define VUINT64x2_ADD_DEFINED +#endif +#ifndef VUINT64x2_SUB_DEFINED +VEC_GENERIC_SUB(u, 64, 2) +# define VUINT64x2_SUB_DEFINED +#endif +#ifndef VUINT64x2_MUL_DEFINED +VEC_GENERIC_MUL(u, 64, 2) +# define VUINT64x2_MUL_DEFINED +#endif +#ifndef VUINT64x2_DIV_DEFINED +VEC_GENERIC_DIV(u, 64, 2) +# define VUINT64x2_DIV_DEFINED +#endif +#ifndef VUINT64x2_AVG_DEFINED +VEC_GENERIC_AVG(u, 64, 2) +# define VUINT64x2_AVG_DEFINED +#endif +#ifndef VUINT64x2_AND_DEFINED +VEC_GENERIC_AND(u, 64, 2) +# define VUINT64x2_AND_DEFINED +#endif +#ifndef VUINT64x2_OR_DEFINED +VEC_GENERIC_OR(u, 64, 2) +# define VUINT64x2_OR_DEFINED +#endif +#ifndef VUINT64x2_XOR_DEFINED +VEC_GENERIC_XOR(u, 64, 2) +# define VUINT64x2_XOR_DEFINED +#endif +#ifndef VUINT64x2_NOT_DEFINED +VEC_GENERIC_NOT(u, 64, 2) +# define VUINT64x2_NOT_DEFINED +#endif +#ifndef VUINT64x2_CMPLT_DEFINED +VEC_GENERIC_CMPLT(u, 64, 2) +# define VUINT64x2_CMPLT_DEFINED +#endif +#ifndef VUINT64x2_CMPEQ_DEFINED +VEC_GENERIC_CMPEQ(u, 64, 2) +# define VUINT64x2_CMPEQ_DEFINED +#endif +#ifndef VUINT64x2_CMPGT_DEFINED +VEC_GENERIC_CMPGT(u, 64, 2) +# define VUINT64x2_CMPGT_DEFINED +#endif +#ifndef VUINT64x2_CMPLE_DEFINED +VEC_GENERIC_CMPLE(u, 64, 2) +# define VUINT64x2_CMPLE_DEFINED +#endif +#ifndef VUINT64x2_CMPGE_DEFINED +VEC_GENERIC_CMPGE(u, 64, 2) +# define VUINT64x2_CMPGE_DEFINED +#endif +#ifndef VUINT64x2_MIN_DEFINED +VEC_GENERIC_MIN(u, 64, 2) +# define VUINT64x2_MIN_DEFINED +#endif +#ifndef VUINT64x2_MAX_DEFINED +VEC_GENERIC_MAX(u, 64, 2) +# define VUINT64x2_MAX_DEFINED +#endif +#ifndef VUINT64x2_RSHIFT_DEFINED +VEC_GENERIC_RSHIFT(u, 64, 2) +# define VUINT64x2_RSHIFT_DEFINED +#endif +#ifndef VUINT64x2_LRSHIFT_DEFINED +VEC_GENERIC_LRSHIFT(u, 64, 2) +# define VUINT64x2_LRSHIFT_DEFINED +#endif +#ifndef VUINT64x2_LSHIFT_DEFINED +VEC_GENERIC_LSHIFT(u, 64, 2) +# define VUINT64x2_LSHIFT_DEFINED +#endif + + +/* vuint64x4 */ + +#ifndef VINT64x4_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(/* nothing */, 64, 4, 2) +# define VINT64x4_SPLAT_DEFINED +#endif + +#ifndef VINT64x4_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(/* nothing */, 64, 4, 2) +# define VINT64x4_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VINT64x4_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(/* nothing */, 64, 4, 2) +# define VINT64x4_LOAD_DEFINED +#endif + +#ifndef VINT64x4_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(/* nothing */, 64, 4, 2) +# define VINT64x4_STORE_ALIGNED_DEFINED +#endif + +#ifndef VINT64x4_STORE_DEFINED +VEC_GENERIC_DBL_STORE(/* nothing */, 64, 4, 2) +# define VINT64x4_STORE_DEFINED +#endif + +#ifndef VINT64x4_ADD_DEFINED +VEC_GENERIC_DBL_ADD(/* nothing */, 64, 4, 2) +# define VINT64x4_ADD_DEFINED +#endif + +#ifndef VINT64x4_SUB_DEFINED +VEC_GENERIC_DBL_SUB(/* nothing */, 64, 4, 2) +# define VINT64x4_SUB_DEFINED +#endif + +#ifndef VINT64x4_MUL_DEFINED +VEC_GENERIC_DBL_MUL(/* nothing */, 64, 4, 2) +# define VINT64x4_MUL_DEFINED +#endif + +#ifndef VINT64x4_DIV_DEFINED +VEC_GENERIC_DBL_DIV(/* nothing */, 64, 4, 2) +# define VINT64x4_DIV_DEFINED +#endif + +#ifndef VINT64x4_AVG_DEFINED +VEC_GENERIC_DBL_AVG(/* nothing */, 64, 4, 2) +# define VINT64x4_AVG_DEFINED +#endif + +#ifndef VINT64x4_AND_DEFINED +VEC_GENERIC_DBL_AND(/* nothing */, 64, 4, 2) +# define VINT64x4_AND_DEFINED +#endif + +#ifndef VINT64x4_OR_DEFINED +VEC_GENERIC_DBL_OR(/* nothing */, 64, 4, 2) +# define VINT64x4_OR_DEFINED +#endif + +#ifndef VINT64x4_XOR_DEFINED +VEC_GENERIC_DBL_XOR(/* nothing */, 64, 4, 2) +# define VINT64x4_XOR_DEFINED +#endif + +#ifndef VINT64x4_NOT_DEFINED +VEC_GENERIC_DBL_NOT(/* nothing */, 64, 4, 2) +# define VINT64x4_NOT_DEFINED +#endif + +#ifndef VINT64x4_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(/* nothing */, 64, 4, 2) +# define VINT64x4_CMPLT_DEFINED +#endif + +#ifndef VINT64x4_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(/* nothing */, 64, 4, 2) +# define VINT64x4_CMPEQ_DEFINED +#endif + +#ifndef VINT64x4_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(/* nothing */, 64, 4, 2) +# define VINT64x4_CMPGT_DEFINED +#endif + +#ifndef VINT64x4_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(/* nothing */, 64, 4, 2) +# define VINT64x4_CMPLE_DEFINED +#endif + +#ifndef VINT64x4_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(/* nothing */, 64, 4, 2) +# define VINT64x4_CMPGE_DEFINED +#endif + +#ifndef VINT64x4_MIN_DEFINED +VEC_GENERIC_DBL_MIN(/* nothing */, 64, 4, 2) +# define VINT64x4_MIN_DEFINED +#endif + +#ifndef VINT64x4_MAX_DEFINED +VEC_GENERIC_DBL_MAX(/* nothing */, 64, 4, 2) +# define VINT64x4_MAX_DEFINED +#endif + +#ifndef VINT64x4_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(/* nothing */, 64, 4, 2) +# define VINT64x4_RSHIFT_DEFINED +#endif + +#ifndef VINT64x4_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(/* nothing */, 64, 4, 2) +# define VINT64x4_LRSHIFT_DEFINED +#endif + +#ifndef VINT64x4_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(/* nothing */, 64, 4, 2) +# define VINT64x4_LSHIFT_DEFINED +#endif + + + +/* vint64x4 */ + +#ifndef VUINT64x4_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(u, 64, 4, 2) +# define VUINT64x4_SPLAT_DEFINED +#endif + +#ifndef VUINT64x4_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(u, 64, 4, 2) +# define VUINT64x4_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VUINT64x4_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(u, 64, 4, 2) +# define VUINT64x4_LOAD_DEFINED +#endif + +#ifndef VUINT64x4_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(u, 64, 4, 2) +# define VUINT64x4_STORE_ALIGNED_DEFINED +#endif + +#ifndef VUINT64x4_STORE_DEFINED +VEC_GENERIC_DBL_STORE(u, 64, 4, 2) +# define VUINT64x4_STORE_DEFINED +#endif + +#ifndef VUINT64x4_ADD_DEFINED +VEC_GENERIC_DBL_ADD(u, 64, 4, 2) +# define VUINT64x4_ADD_DEFINED +#endif + +#ifndef VUINT64x4_SUB_DEFINED +VEC_GENERIC_DBL_SUB(u, 64, 4, 2) +# define VUINT64x4_SUB_DEFINED +#endif + +#ifndef VUINT64x4_MUL_DEFINED +VEC_GENERIC_DBL_MUL(u, 64, 4, 2) +# define VUINT64x4_MUL_DEFINED +#endif + +#ifndef VUINT64x4_DIV_DEFINED +VEC_GENERIC_DBL_DIV(u, 64, 4, 2) +# define VUINT64x4_DIV_DEFINED +#endif + +#ifndef VUINT64x4_AVG_DEFINED +VEC_GENERIC_DBL_AVG(u, 64, 4, 2) +# define VUINT64x4_AVG_DEFINED +#endif + +#ifndef VUINT64x4_AND_DEFINED +VEC_GENERIC_DBL_AND(u, 64, 4, 2) +# define VUINT64x4_AND_DEFINED +#endif + +#ifndef VUINT64x4_OR_DEFINED +VEC_GENERIC_DBL_OR(u, 64, 4, 2) +# define VUINT64x4_OR_DEFINED +#endif + +#ifndef VUINT64x4_XOR_DEFINED +VEC_GENERIC_DBL_XOR(u, 64, 4, 2) +# define VUINT64x4_XOR_DEFINED +#endif + +#ifndef VUINT64x4_NOT_DEFINED +VEC_GENERIC_DBL_NOT(u, 64, 4, 2) +# define VUINT64x4_NOT_DEFINED +#endif + +#ifndef VUINT64x4_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(u, 64, 4, 2) +# define VUINT64x4_CMPLT_DEFINED +#endif + +#ifndef VUINT64x4_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(u, 64, 4, 2) +# define VUINT64x4_CMPEQ_DEFINED +#endif + +#ifndef VUINT64x4_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(u, 64, 4, 2) +# define VUINT64x4_CMPGT_DEFINED +#endif + +#ifndef VUINT64x4_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(u, 64, 4, 2) +# define VUINT64x4_CMPLE_DEFINED +#endif + +#ifndef VUINT64x4_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(u, 64, 4, 2) +# define VUINT64x4_CMPGE_DEFINED +#endif + +#ifndef VUINT64x4_MIN_DEFINED +VEC_GENERIC_DBL_MIN(u, 64, 4, 2) +# define VUINT64x4_MIN_DEFINED +#endif + +#ifndef VUINT64x4_MAX_DEFINED +VEC_GENERIC_DBL_MAX(u, 64, 4, 2) +# define VUINT64x4_MAX_DEFINED +#endif + +#ifndef VUINT64x4_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(u, 64, 4, 2) +# define VUINT64x4_RSHIFT_DEFINED +#endif + +#ifndef VUINT64x4_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(u, 64, 4, 2) +# define VUINT64x4_LRSHIFT_DEFINED +#endif + +#ifndef VUINT64x4_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(u, 64, 4, 2) +# define VUINT64x4_LSHIFT_DEFINED +#endif + + + +/* vuint64x8 */ + +#ifndef VINT64x8_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(/* nothing */, 64, 8, 4) +# define VINT64x8_SPLAT_DEFINED +#endif + +#ifndef VINT64x8_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(/* nothing */, 64, 8, 4) +# define VINT64x8_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VINT64x8_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(/* nothing */, 64, 8, 4) +# define VINT64x8_LOAD_DEFINED +#endif + +#ifndef VINT64x8_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(/* nothing */, 64, 8, 4) +# define VINT64x8_STORE_ALIGNED_DEFINED +#endif + +#ifndef VINT64x8_STORE_DEFINED +VEC_GENERIC_DBL_STORE(/* nothing */, 64, 8, 4) +# define VINT64x8_STORE_DEFINED +#endif + +#ifndef VINT64x8_ADD_DEFINED +VEC_GENERIC_DBL_ADD(/* nothing */, 64, 8, 4) +# define VINT64x8_ADD_DEFINED +#endif + +#ifndef VINT64x8_SUB_DEFINED +VEC_GENERIC_DBL_SUB(/* nothing */, 64, 8, 4) +# define VINT64x8_SUB_DEFINED +#endif + +#ifndef VINT64x8_MUL_DEFINED +VEC_GENERIC_DBL_MUL(/* nothing */, 64, 8, 4) +# define VINT64x8_MUL_DEFINED +#endif + +#ifndef VINT64x8_DIV_DEFINED +VEC_GENERIC_DBL_DIV(/* nothing */, 64, 8, 4) +# define VINT64x8_DIV_DEFINED +#endif + +#ifndef VINT64x8_AVG_DEFINED +VEC_GENERIC_DBL_AVG(/* nothing */, 64, 8, 4) +# define VINT64x8_AVG_DEFINED +#endif + +#ifndef VINT64x8_AND_DEFINED +VEC_GENERIC_DBL_AND(/* nothing */, 64, 8, 4) +# define VINT64x8_AND_DEFINED +#endif + +#ifndef VINT64x8_OR_DEFINED +VEC_GENERIC_DBL_OR(/* nothing */, 64, 8, 4) +# define VINT64x8_OR_DEFINED +#endif + +#ifndef VINT64x8_XOR_DEFINED +VEC_GENERIC_DBL_XOR(/* nothing */, 64, 8, 4) +# define VINT64x8_XOR_DEFINED +#endif + +#ifndef VINT64x8_NOT_DEFINED +VEC_GENERIC_DBL_NOT(/* nothing */, 64, 8, 4) +# define VINT64x8_NOT_DEFINED +#endif + +#ifndef VINT64x8_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(/* nothing */, 64, 8, 4) +# define VINT64x8_CMPLT_DEFINED +#endif + +#ifndef VINT64x8_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(/* nothing */, 64, 8, 4) +# define VINT64x8_CMPEQ_DEFINED +#endif + +#ifndef VINT64x8_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(/* nothing */, 64, 8, 4) +# define VINT64x8_CMPGT_DEFINED +#endif + +#ifndef VINT64x8_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(/* nothing */, 64, 8, 4) +# define VINT64x8_CMPLE_DEFINED +#endif + +#ifndef VINT64x8_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(/* nothing */, 64, 8, 4) +# define VINT64x8_CMPGE_DEFINED +#endif + +#ifndef VINT64x8_MIN_DEFINED +VEC_GENERIC_DBL_MIN(/* nothing */, 64, 8, 4) +# define VINT64x8_MIN_DEFINED +#endif + +#ifndef VINT64x8_MAX_DEFINED +VEC_GENERIC_DBL_MAX(/* nothing */, 64, 8, 4) +# define VINT64x8_MAX_DEFINED +#endif + +#ifndef VINT64x8_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(/* nothing */, 64, 8, 4) +# define VINT64x8_RSHIFT_DEFINED +#endif + +#ifndef VINT64x8_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(/* nothing */, 64, 8, 4) +# define VINT64x8_LRSHIFT_DEFINED +#endif + +#ifndef VINT64x8_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(/* nothing */, 64, 8, 4) +# define VINT64x8_LSHIFT_DEFINED +#endif + + + +/* vint64x8 */ + +#ifndef VUINT64x8_SPLAT_DEFINED +VEC_GENERIC_DBL_SPLAT(u, 64, 8, 4) +# define VUINT64x8_SPLAT_DEFINED +#endif + +#ifndef VUINT64x8_LOAD_ALIGNED_DEFINED +VEC_GENERIC_DBL_LOAD_ALIGNED(u, 64, 8, 4) +# define VUINT64x8_LOAD_ALIGNED_DEFINED +#endif + +#ifndef VUINT64x8_LOAD_DEFINED +VEC_GENERIC_DBL_LOAD(u, 64, 8, 4) +# define VUINT64x8_LOAD_DEFINED +#endif + +#ifndef VUINT64x8_STORE_ALIGNED_DEFINED +VEC_GENERIC_DBL_STORE_ALIGNED(u, 64, 8, 4) +# define VUINT64x8_STORE_ALIGNED_DEFINED +#endif + +#ifndef VUINT64x8_STORE_DEFINED +VEC_GENERIC_DBL_STORE(u, 64, 8, 4) +# define VUINT64x8_STORE_DEFINED +#endif + +#ifndef VUINT64x8_ADD_DEFINED +VEC_GENERIC_DBL_ADD(u, 64, 8, 4) +# define VUINT64x8_ADD_DEFINED +#endif + +#ifndef VUINT64x8_SUB_DEFINED +VEC_GENERIC_DBL_SUB(u, 64, 8, 4) +# define VUINT64x8_SUB_DEFINED +#endif + +#ifndef VUINT64x8_MUL_DEFINED +VEC_GENERIC_DBL_MUL(u, 64, 8, 4) +# define VUINT64x8_MUL_DEFINED +#endif + +#ifndef VUINT64x8_DIV_DEFINED +VEC_GENERIC_DBL_DIV(u, 64, 8, 4) +# define VUINT64x8_DIV_DEFINED +#endif + +#ifndef VUINT64x8_AVG_DEFINED +VEC_GENERIC_DBL_AVG(u, 64, 8, 4) +# define VUINT64x8_AVG_DEFINED +#endif + +#ifndef VUINT64x8_AND_DEFINED +VEC_GENERIC_DBL_AND(u, 64, 8, 4) +# define VUINT64x8_AND_DEFINED +#endif + +#ifndef VUINT64x8_OR_DEFINED +VEC_GENERIC_DBL_OR(u, 64, 8, 4) +# define VUINT64x8_OR_DEFINED +#endif + +#ifndef VUINT64x8_XOR_DEFINED +VEC_GENERIC_DBL_XOR(u, 64, 8, 4) +# define VUINT64x8_XOR_DEFINED +#endif + +#ifndef VUINT64x8_NOT_DEFINED +VEC_GENERIC_DBL_NOT(u, 64, 8, 4) +# define VUINT64x8_NOT_DEFINED +#endif + +#ifndef VUINT64x8_CMPLT_DEFINED +VEC_GENERIC_DBL_CMPLT(u, 64, 8, 4) +# define VUINT64x8_CMPLT_DEFINED +#endif + +#ifndef VUINT64x8_CMPEQ_DEFINED +VEC_GENERIC_DBL_CMPEQ(u, 64, 8, 4) +# define VUINT64x8_CMPEQ_DEFINED +#endif + +#ifndef VUINT64x8_CMPGT_DEFINED +VEC_GENERIC_DBL_CMPGT(u, 64, 8, 4) +# define VUINT64x8_CMPGT_DEFINED +#endif + +#ifndef VUINT64x8_CMPLE_DEFINED +VEC_GENERIC_DBL_CMPLE(u, 64, 8, 4) +# define VUINT64x8_CMPLE_DEFINED +#endif + +#ifndef VUINT64x8_CMPGE_DEFINED +VEC_GENERIC_DBL_CMPGE(u, 64, 8, 4) +# define VUINT64x8_CMPGE_DEFINED +#endif + +#ifndef VUINT64x8_MIN_DEFINED +VEC_GENERIC_DBL_MIN(u, 64, 8, 4) +# define VUINT64x8_MIN_DEFINED +#endif + +#ifndef VUINT64x8_MAX_DEFINED +VEC_GENERIC_DBL_MAX(u, 64, 8, 4) +# define VUINT64x8_MAX_DEFINED +#endif + +#ifndef VUINT64x8_RSHIFT_DEFINED +VEC_GENERIC_DBL_RSHIFT(u, 64, 8, 4) +# define VUINT64x8_RSHIFT_DEFINED +#endif + +#ifndef VUINT64x8_LRSHIFT_DEFINED +VEC_GENERIC_DBL_LRSHIFT(u, 64, 8, 4) +# define VUINT64x8_LRSHIFT_DEFINED +#endif + +#ifndef VUINT64x8_LSHIFT_DEFINED +VEC_GENERIC_DBL_LSHIFT(u, 64, 8, 4) +# define VUINT64x8_LSHIFT_DEFINED +#endif #endif /* VEC_IMPL_GENERIC_H_ */ +